[ 476.884432] env[62000]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62000) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 476.884761] env[62000]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62000) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 476.884854] env[62000]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62000) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 476.885144] env[62000]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 476.977862] env[62000]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62000) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 476.987567] env[62000]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62000) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 477.590099] env[62000]: INFO nova.virt.driver [None req-43f896ad-6727-4c0e-8e28-f20ed24a52c0 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 477.660289] env[62000]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 477.660483] env[62000]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 477.660555] env[62000]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62000) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 480.766784] env[62000]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-2ea45181-5681-4165-8118-4dbee4ba6e4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.783060] env[62000]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62000) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 480.783249] env[62000]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-7f7fefbe-ac3f-4434-91de-416a76702035 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.819538] env[62000]: INFO oslo_vmware.api [-] Successfully established new session; session ID is ace3f. [ 480.819781] env[62000]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.159s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 480.820305] env[62000]: INFO nova.virt.vmwareapi.driver [None req-43f896ad-6727-4c0e-8e28-f20ed24a52c0 None None] VMware vCenter version: 7.0.3 [ 480.823737] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75f24b7c-07c3-4e42-8912-628707770b92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.841167] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8a4d85e-ea46-4ab2-8382-2e5c54e0a852 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.847165] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-423ab990-cfd0-4fcf-897d-43570184dcdf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.853810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3ba8b54-1889-4cb1-9b60-99d9e70ca5cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.866857] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25a7d1f-ef70-4f61-ac58-e4bbb0a7b5d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.872604] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-091db840-445c-4c53-a42f-1e5e1c6b56db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.902208] env[62000]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-0067e60b-9a5b-4642-a45d-800d2cc7d59d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 480.906924] env[62000]: DEBUG nova.virt.vmwareapi.driver [None req-43f896ad-6727-4c0e-8e28-f20ed24a52c0 None None] Extension org.openstack.compute already exists. {{(pid=62000) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 480.909591] env[62000]: INFO nova.compute.provider_config [None req-43f896ad-6727-4c0e-8e28-f20ed24a52c0 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 481.413176] env[62000]: DEBUG nova.context [None req-43f896ad-6727-4c0e-8e28-f20ed24a52c0 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),766c0c9c-9c46-477d-aa04-1828e62012f3(cell1) {{(pid=62000) load_cells /opt/stack/nova/nova/context.py:464}} [ 481.415471] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.415750] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 481.416463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 481.416929] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Acquiring lock "766c0c9c-9c46-477d-aa04-1828e62012f3" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 481.417234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Lock "766c0c9c-9c46-477d-aa04-1828e62012f3" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 481.418537] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Lock "766c0c9c-9c46-477d-aa04-1828e62012f3" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 481.439471] env[62000]: INFO dbcounter [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Registered counter for database nova_cell0 [ 481.448544] env[62000]: INFO dbcounter [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Registered counter for database nova_cell1 [ 481.452073] env[62000]: DEBUG oslo_db.sqlalchemy.engines [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62000) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 481.452444] env[62000]: DEBUG oslo_db.sqlalchemy.engines [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62000) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 481.457413] env[62000]: ERROR nova.db.main.api [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 481.457413] env[62000]: result = function(*args, **kwargs) [ 481.457413] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 481.457413] env[62000]: return func(*args, **kwargs) [ 481.457413] env[62000]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 481.457413] env[62000]: result = fn(*args, **kwargs) [ 481.457413] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 481.457413] env[62000]: return f(*args, **kwargs) [ 481.457413] env[62000]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 481.457413] env[62000]: return db.service_get_minimum_version(context, binaries) [ 481.457413] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 481.457413] env[62000]: _check_db_access() [ 481.457413] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 481.457413] env[62000]: stacktrace = ''.join(traceback.format_stack()) [ 481.457413] env[62000]: [ 481.458225] env[62000]: ERROR nova.db.main.api [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 481.458225] env[62000]: result = function(*args, **kwargs) [ 481.458225] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 481.458225] env[62000]: return func(*args, **kwargs) [ 481.458225] env[62000]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 481.458225] env[62000]: result = fn(*args, **kwargs) [ 481.458225] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 481.458225] env[62000]: return f(*args, **kwargs) [ 481.458225] env[62000]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 481.458225] env[62000]: return db.service_get_minimum_version(context, binaries) [ 481.458225] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 481.458225] env[62000]: _check_db_access() [ 481.458225] env[62000]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 481.458225] env[62000]: stacktrace = ''.join(traceback.format_stack()) [ 481.458225] env[62000]: [ 481.458628] env[62000]: WARNING nova.objects.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 481.458739] env[62000]: WARNING nova.objects.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Failed to get minimum service version for cell 766c0c9c-9c46-477d-aa04-1828e62012f3 [ 481.459172] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Acquiring lock "singleton_lock" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 481.459337] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Acquired lock "singleton_lock" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 481.459581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Releasing lock "singleton_lock" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 481.459931] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Full set of CONF: {{(pid=62000) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 481.460094] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ******************************************************************************** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 481.460224] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Configuration options gathered from: {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 481.460360] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 481.460554] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 481.460684] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ================================================================================ {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 481.460897] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] allow_resize_to_same_host = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461083] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] arq_binding_timeout = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461305] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] backdoor_port = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461450] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] backdoor_socket = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461511] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] block_device_allocate_retries = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461667] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] block_device_allocate_retries_interval = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.461859] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cert = self.pem {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462054] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462237] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute_monitors = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462409] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] config_dir = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462586] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] config_drive_format = iso9660 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462723] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.462915] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] config_source = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463109] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] console_host = devstack {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463282] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] control_exchange = nova {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463442] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cpu_allocation_ratio = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463602] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] daemon = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463768] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] debug = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.463927] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_access_ip_network_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464111] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_availability_zone = nova {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464272] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_ephemeral_format = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464432] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_green_pool_size = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464670] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464834] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] default_schedule_zone = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.464993] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] disk_allocation_ratio = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465168] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] enable_new_services = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465346] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] enabled_apis = ['osapi_compute'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465510] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] enabled_ssl_apis = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465670] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] flat_injected = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465827] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] force_config_drive = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.465986] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] force_raw_images = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.466167] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] graceful_shutdown_timeout = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.466330] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] heal_instance_info_cache_interval = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.466538] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] host = cpu-1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.466713] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.466877] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] initial_disk_allocation_ratio = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.467048] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] initial_ram_allocation_ratio = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.467275] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.467471] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_build_timeout = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.467697] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_delete_interval = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.467881] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_format = [instance: %(uuid)s] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468070] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_name_template = instance-%08x {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468248] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_usage_audit = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468414] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_usage_audit_period = month {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468580] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468744] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] instances_path = /opt/stack/data/nova/instances {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.468907] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] internal_service_availability_zone = internal {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469077] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] key = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469242] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] live_migration_retry_count = 30 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469410] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_color = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469573] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_config_append = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469739] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.469899] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_dir = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470068] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470198] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_options = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470359] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_rotate_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470526] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_rotate_interval_type = days {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470690] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] log_rotation_type = none {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470824] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.470952] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471132] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471295] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471421] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471583] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] long_rpc_timeout = 1800 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471741] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_concurrent_builds = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.471925] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_concurrent_live_migrations = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472104] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_concurrent_snapshots = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472265] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_local_block_devices = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472423] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_logfile_count = 30 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472578] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] max_logfile_size_mb = 200 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472734] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] maximum_instance_delete_attempts = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.472922] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metadata_listen = 0.0.0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473108] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metadata_listen_port = 8775 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473278] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metadata_workers = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473438] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] migrate_max_retries = -1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473602] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] mkisofs_cmd = genisoimage {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473807] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] my_block_storage_ip = 10.180.1.21 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.473939] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] my_ip = 10.180.1.21 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474112] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] network_allocate_retries = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474291] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474455] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] osapi_compute_listen = 0.0.0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474616] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] osapi_compute_listen_port = 8774 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474780] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] osapi_compute_unique_server_name_scope = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.474947] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] osapi_compute_workers = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475125] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] password_length = 12 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475288] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] periodic_enable = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475447] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] periodic_fuzzy_delay = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475613] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] pointer_model = usbtablet {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475777] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] preallocate_images = none {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.475933] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] publish_errors = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476071] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] pybasedir = /opt/stack/nova {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476231] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ram_allocation_ratio = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476388] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rate_limit_burst = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476568] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rate_limit_except_level = CRITICAL {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476710] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rate_limit_interval = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.476865] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reboot_timeout = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477034] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reclaim_instance_interval = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477194] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] record = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477359] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reimage_timeout_per_gb = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477522] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] report_interval = 120 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477682] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rescue_timeout = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477842] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reserved_host_cpus = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.477999] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reserved_host_disk_mb = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478172] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reserved_host_memory_mb = 512 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478329] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] reserved_huge_pages = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478489] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] resize_confirm_window = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478645] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] resize_fs_using_block_device = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478801] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] resume_guests_state_on_host_boot = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.478966] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479139] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] rpc_response_timeout = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479300] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] run_external_periodic_tasks = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479466] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] running_deleted_instance_action = reap {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479626] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] running_deleted_instance_poll_interval = 1800 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479784] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] running_deleted_instance_timeout = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.479943] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler_instance_sync_interval = 120 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480122] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_down_time = 720 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480292] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] servicegroup_driver = db {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480447] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] shell_completion = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480606] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] shelved_offload_time = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480767] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] shelved_poll_interval = 3600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.480935] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] shutdown_timeout = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.481109] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] source_is_ipv6 = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.481269] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ssl_only = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.481514] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.481682] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] sync_power_state_interval = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.481863] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] sync_power_state_pool_size = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482056] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] syslog_log_facility = LOG_USER {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482217] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] tempdir = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482377] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] timeout_nbd = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482543] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] transport_url = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482703] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] update_resources_interval = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.482897] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_cow_images = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483076] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_eventlog = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483238] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_journal = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483397] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_json = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483554] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_rootwrap_daemon = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483709] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_stderr = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.483864] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] use_syslog = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484028] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vcpu_pin_set = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484200] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plugging_is_fatal = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484366] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plugging_timeout = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484530] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] virt_mkfs = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484690] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] volume_usage_poll_interval = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.484848] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] watch_log_file = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.485026] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] web = /usr/share/spice-html5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 481.485208] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_concurrency.disable_process_locking = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.485785] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.485982] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.486175] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.486356] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.486543] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.486715] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.486899] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.auth_strategy = keystone {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487081] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.compute_link_prefix = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487265] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487441] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.dhcp_domain = novalocal {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487612] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.enable_instance_password = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487783] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.glance_link_prefix = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.487952] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488142] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488309] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.instance_list_per_project_cells = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488476] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.list_records_by_skipping_down_cells = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488640] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.local_metadata_per_cell = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488812] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.max_limit = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.488984] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.metadata_cache_expiration = 15 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.489174] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.neutron_default_tenant_id = default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.489347] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.response_validation = warn {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.489517] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.use_neutron_default_nets = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.489688] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.489853] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490037] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490219] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490394] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_dynamic_targets = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490559] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_jsonfile_path = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490743] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.490939] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.backend = dogpile.cache.memcached {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.491130] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.backend_argument = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.491307] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.config_prefix = cache.oslo {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.491479] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.dead_timeout = 60.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.491646] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.debug_cache_backend = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.491826] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.enable_retry_client = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492028] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.enable_socket_keepalive = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492211] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.enabled = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492379] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.enforce_fips_mode = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492544] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.expiration_time = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492709] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.hashclient_retry_attempts = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.492904] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.hashclient_retry_delay = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493096] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_dead_retry = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493263] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_password = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493430] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493595] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493758] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_pool_maxsize = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.493923] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494100] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_sasl_enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494286] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494455] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_socket_timeout = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494617] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.memcache_username = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494785] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.proxies = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.494950] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_db = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495128] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_password = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495300] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_sentinel_service_name = mymaster {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495475] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495653] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_server = localhost:6379 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495822] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_socket_timeout = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.495984] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.redis_username = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496162] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.retry_attempts = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496327] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.retry_delay = 0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496490] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.socket_keepalive_count = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496652] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.socket_keepalive_idle = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496816] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.socket_keepalive_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.496977] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.tls_allowed_ciphers = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497152] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.tls_cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497312] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.tls_certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497475] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.tls_enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497633] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cache.tls_keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497800] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.497975] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.auth_type = password {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498154] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498330] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.catalog_info = volumev3::publicURL {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498492] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498656] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498818] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.cross_az_attach = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.498982] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.debug = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499159] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.endpoint_template = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499324] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.http_retries = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499486] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499641] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499813] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.os_region_name = RegionOne {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.499977] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500152] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cinder.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500323] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500485] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.cpu_dedicated_set = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500644] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.cpu_shared_set = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500811] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.image_type_exclude_list = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.500979] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.501158] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.max_concurrent_disk_ops = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.501323] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.max_disk_devices_to_attach = -1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.501486] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.501653] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.501854] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.resource_provider_association_refresh = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502020] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502191] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.shutdown_retry_interval = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502379] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502558] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] conductor.workers = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502736] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] console.allowed_origins = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.502941] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] console.ssl_ciphers = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503147] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] console.ssl_minimum_version = default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503325] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] consoleauth.enforce_session_timeout = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503496] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] consoleauth.token_ttl = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503665] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503825] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.503989] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504167] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504327] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504487] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504649] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504808] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.504970] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505142] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505303] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505462] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505620] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505788] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.service_type = accelerator {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.505951] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506125] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506286] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506445] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506625] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506789] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] cyborg.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.506969] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.backend = sqlalchemy {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507153] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.connection = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507321] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.connection_debug = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507490] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.connection_parameters = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507654] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.connection_recycle_time = 3600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507816] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.connection_trace = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.507979] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.db_inc_retry_interval = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508157] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.db_max_retries = 20 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508321] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.db_max_retry_interval = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508483] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.db_retry_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508645] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.max_overflow = 50 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508811] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.max_pool_size = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.508974] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.max_retries = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509156] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509316] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.mysql_wsrep_sync_wait = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509476] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.pool_timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509639] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.retry_interval = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509798] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.slave_connection = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.509962] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.sqlite_synchronous = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510137] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] database.use_db_reconnect = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510316] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.backend = sqlalchemy {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510486] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.connection = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510650] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.connection_debug = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510818] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.connection_parameters = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.510983] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.connection_recycle_time = 3600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.511159] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.connection_trace = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.511321] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.db_inc_retry_interval = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.511484] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.db_max_retries = 20 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.511646] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.db_max_retry_interval = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.511822] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.db_retry_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512042] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.max_overflow = 50 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512218] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.max_pool_size = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512385] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.max_retries = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512557] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512719] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.512917] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.pool_timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513116] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.retry_interval = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513285] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.slave_connection = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513450] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] api_database.sqlite_synchronous = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513628] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] devices.enabled_mdev_types = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513809] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.513981] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ephemeral_storage_encryption.default_format = luks {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514161] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ephemeral_storage_encryption.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514327] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514498] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.api_servers = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514664] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514827] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.514992] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.515166] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.515326] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.515489] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.debug = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.515656] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.default_trusted_certificate_ids = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.515821] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.enable_certificate_validation = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516037] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.enable_rbd_download = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516222] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516392] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516557] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516720] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.516881] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517062] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.num_retries = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517238] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.rbd_ceph_conf = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517404] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.rbd_connect_timeout = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517573] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.rbd_pool = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517741] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.rbd_user = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.517906] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518118] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518291] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518463] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.service_type = image {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518626] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518787] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.518947] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519122] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519315] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519479] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.verify_glance_signatures = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519639] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] glance.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519807] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] guestfs.debug = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.519976] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] mks.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.520365] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.520569] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.manager_interval = 2400 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.520752] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.precache_concurrency = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.520928] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.remove_unused_base_images = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.521118] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.521295] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.521473] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] image_cache.subdirectory_name = _base {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.521649] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.api_max_retries = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.521828] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.api_retry_interval = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522053] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522199] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.auth_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522361] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522522] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522686] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.522888] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.conductor_group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523084] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523252] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523413] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523578] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523735] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.523902] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524100] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524278] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.peer_list = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524439] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524597] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524764] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.serial_console_state_timeout = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.524924] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525107] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.service_type = baremetal {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525269] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.shard = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525432] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525623] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525784] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.525945] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.526144] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.526309] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ironic.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.526492] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.526665] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] key_manager.fixed_key = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.526847] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527061] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.barbican_api_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527187] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.barbican_endpoint = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527356] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.barbican_endpoint_type = public {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527514] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.barbican_region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527671] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527828] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.527993] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528169] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528327] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528493] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.number_of_retries = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528654] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.retry_delay = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528816] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.send_service_user_token = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.528980] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529151] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529312] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.verify_ssl = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529469] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican.verify_ssl_path = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529635] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529810] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.auth_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.529996] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530175] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530338] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530503] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530661] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530826] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.530987] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] barbican_service_user.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.531171] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.approle_role_id = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.531334] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.approle_secret_id = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.531501] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.kv_mountpoint = secret {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.531661] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.kv_path = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.531845] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.kv_version = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532035] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.namespace = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532206] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.root_token_id = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532367] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.ssl_ca_crt_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532536] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.timeout = 60.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532697] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.use_ssl = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.532895] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533087] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533260] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.auth_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533420] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533582] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533745] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.533905] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534079] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534242] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534405] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534563] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534722] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.534879] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535052] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535213] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535373] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535564] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.service_type = identity {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535740] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.535902] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536111] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536282] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536466] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536627] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] keystone.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536828] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.connection_uri = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.536992] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_mode = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.537175] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_model_extra_flags = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.537344] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_models = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.537514] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_power_governor_high = performance {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.537683] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_power_governor_low = powersave {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.537846] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_power_management = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538026] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538200] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.device_detach_attempts = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538364] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.device_detach_timeout = 20 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538529] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.disk_cachemodes = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538686] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.disk_prefix = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.538851] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.enabled_perf_events = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539024] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.file_backed_memory = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539196] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.gid_maps = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539355] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.hw_disk_discard = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539512] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.hw_machine_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539678] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_rbd_ceph_conf = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.539857] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540050] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540231] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_rbd_glance_store_name = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540407] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_rbd_pool = rbd {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540580] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_type = default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540742] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.images_volume_group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.540904] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.inject_key = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541081] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.inject_partition = -2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541245] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.inject_password = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541409] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.iscsi_iface = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541568] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.iser_use_multipath = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541730] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_bandwidth = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.541935] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542131] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_downtime = 500 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542302] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542464] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542627] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_inbound_addr = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542794] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.542958] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_permit_post_copy = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543135] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_scheme = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543310] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_timeout_action = abort {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543478] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_tunnelled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543643] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_uri = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543808] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.live_migration_with_native_tls = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.543972] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.max_queues = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.544186] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.544434] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.544603] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.nfs_mount_options = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.544892] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545079] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545248] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_iser_scan_tries = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545410] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_memory_encrypted_guests = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545574] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545735] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_pcie_ports = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.545902] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.num_volume_scan_tries = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.546079] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.pmem_namespaces = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.546242] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.quobyte_client_cfg = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.546522] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.546694] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rbd_connect_timeout = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.546862] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547037] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547209] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rbd_secret_uuid = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547365] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rbd_user = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547528] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547696] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.remote_filesystem_transport = ssh {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.547857] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rescue_image_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.548052] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rescue_kernel_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.548235] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rescue_ramdisk_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.548406] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.548565] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.rx_queue_size = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.548731] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.smbfs_mount_options = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549014] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549202] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.snapshot_compression = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549367] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.snapshot_image_format = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549586] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549752] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.sparse_logical_volumes = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.549915] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.swtpm_enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550098] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.swtpm_group = tss {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550269] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.swtpm_user = tss {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550437] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.sysinfo_serial = unique {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550596] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.tb_cache_size = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550753] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.tx_queue_size = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.550918] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.uid_maps = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551093] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.use_virtio_for_bridges = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551266] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.virt_type = kvm {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551436] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.volume_clear = zero {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551597] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.volume_clear_size = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551763] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.volume_use_multipath = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.551948] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_cache_path = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.552146] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.552333] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_mount_group = qemu {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.552483] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_mount_opts = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.552649] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.552981] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.553208] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.vzstorage_mount_user = stack {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.553421] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.553625] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.553907] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.auth_type = password {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554123] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554298] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554470] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554634] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554795] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.554994] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.default_floating_pool = public {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.555191] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.555411] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.extension_sync_interval = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.555594] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.http_retries = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.555770] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.555977] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.556164] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.556341] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.556501] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.556670] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.ovs_bridge = br-int {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.556840] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.physnets = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557015] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.region_name = RegionOne {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557183] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557393] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.service_metadata_proxy = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557542] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557711] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.service_type = network {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.557873] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558071] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558207] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558364] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558546] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558738] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] neutron.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.558929] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] notifications.bdms_in_notifications = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.559189] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] notifications.default_level = INFO {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.559397] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] notifications.notification_format = unversioned {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.559566] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] notifications.notify_on_state_change = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.559745] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.559955] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] pci.alias = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.560155] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] pci.device_spec = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.560324] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] pci.report_in_placement = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.560498] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.560672] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.auth_type = password {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.560838] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561006] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561174] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561337] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561497] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561653] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.561824] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.default_domain_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562018] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.default_domain_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562189] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.domain_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562349] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.domain_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562509] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562670] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.562845] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563025] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563187] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563355] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.password = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563512] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.project_domain_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563676] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.project_domain_name = Default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.563843] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.project_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564049] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.project_name = service {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564232] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.region_name = RegionOne {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564397] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564556] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564726] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.service_type = placement {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.564886] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565062] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565230] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565391] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.system_scope = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565549] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565705] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.trust_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.565863] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.user_domain_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566044] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.user_domain_name = Default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566207] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.user_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566379] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.username = nova {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566560] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566719] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] placement.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.566899] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.cores = 20 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567075] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.count_usage_from_placement = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567248] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567495] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.injected_file_content_bytes = 10240 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567620] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.injected_file_path_length = 255 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567790] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.injected_files = 5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.567959] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.instances = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568140] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.key_pairs = 100 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568310] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.metadata_items = 128 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568475] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.ram = 51200 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568638] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.recheck_quota = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568806] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.server_group_members = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.568973] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] quota.server_groups = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569163] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569336] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569499] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.image_metadata_prefilter = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569659] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569828] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.max_attempts = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.569987] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.max_placement_results = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.570166] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.570330] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.query_placement_for_image_type_support = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.570490] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.570664] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] scheduler.workers = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.570837] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571015] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571203] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571371] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571536] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571699] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.571867] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572069] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572244] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.host_subset_size = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572408] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572568] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572729] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.572917] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.isolated_hosts = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.573163] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.isolated_images = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.573350] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.573515] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.573680] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.573842] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.pci_in_placement = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574014] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574184] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574351] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574513] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574674] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574835] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.574996] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.track_instance_changes = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.575187] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.575356] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metrics.required = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.575514] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metrics.weight_multiplier = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.575674] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.575837] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] metrics.weight_setting = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.576169] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.576344] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.576526] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.port_range = 10000:20000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.576697] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.576866] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577042] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] serial_console.serialproxy_port = 6083 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577213] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577384] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.auth_type = password {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577546] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577696] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.577854] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578023] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578185] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578355] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.send_service_user_token = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578515] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578671] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] service_user.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.578839] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.agent_enabled = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.579007] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.579334] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.579529] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.579702] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.html5proxy_port = 6082 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.579885] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.image_compression = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580073] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.jpeg_compression = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580240] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.playback_compression = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580405] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.require_secure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580576] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.server_listen = 127.0.0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580746] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.580906] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.streaming_mode = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581078] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] spice.zlib_compression = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581249] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] upgrade_levels.baseapi = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581423] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] upgrade_levels.compute = auto {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581583] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] upgrade_levels.conductor = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581741] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] upgrade_levels.scheduler = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.581935] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582119] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.auth_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582284] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582442] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582632] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582759] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.582950] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583123] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583285] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vendordata_dynamic_auth.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583459] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.api_retry_count = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583620] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.ca_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583809] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.cache_prefix = devstack-image-cache {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.583999] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.cluster_name = testcl1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.584179] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.connection_pool_size = 10 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.584341] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.console_delay_seconds = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.584508] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.datastore_regex = ^datastore.* {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.584720] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.584891] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.host_password = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585070] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.host_port = 443 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585244] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.host_username = administrator@vsphere.local {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585410] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.insecure = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585572] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.integration_bridge = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585736] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.maximum_objects = 100 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.585896] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.pbm_default_policy = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586069] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.pbm_enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586231] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.pbm_wsdl_location = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586398] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586556] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.serial_port_proxy_uri = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586714] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.serial_port_service_uri = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.586880] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.task_poll_interval = 0.5 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.587065] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.use_linked_clone = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.587280] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.vnc_keymap = en-us {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.587515] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.vnc_port = 5900 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.587739] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vmware.vnc_port_total = 10000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.587886] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.auth_schemes = ['none'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.588076] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.588381] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.588567] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.588748] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.novncproxy_port = 6080 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.588970] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.server_listen = 127.0.0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.589171] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.589337] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.vencrypt_ca_certs = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.589500] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.vencrypt_client_cert = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.589658] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vnc.vencrypt_client_key = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.589864] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590052] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_deep_image_inspection = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590221] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590386] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590550] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590713] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.disable_rootwrap = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.590876] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.enable_numa_live_migration = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591051] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591214] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591377] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591538] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.libvirt_disable_apic = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591698] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.591884] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592069] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592236] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592400] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592560] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592735] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.592882] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593053] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593227] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593410] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593580] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.client_socket_timeout = 900 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593746] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.default_pool_size = 1000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.593912] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.keep_alive = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594094] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.max_header_line = 16384 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594260] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.secure_proxy_ssl_header = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594422] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.ssl_ca_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594582] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.ssl_cert_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594740] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.ssl_key_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.594908] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.tcp_keepidle = 600 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.595099] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.595271] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] zvm.ca_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.595434] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] zvm.cloud_connector_url = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.595731] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.595930] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] zvm.reachable_timeout = 300 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.596136] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.enforce_new_defaults = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.596551] env[62000]: WARNING oslo_config.cfg [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 481.596739] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.enforce_scope = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.596920] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.policy_default_rule = default {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597121] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597303] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.policy_file = policy.yaml {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597480] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597642] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597840] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.597961] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.598143] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.598315] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.598492] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.598668] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.connection_string = messaging:// {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.598835] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.enabled = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599012] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.es_doc_type = notification {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599185] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.es_scroll_size = 10000 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599353] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.es_scroll_time = 2m {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599516] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.filter_error_trace = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599688] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.hmac_keys = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.599877] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.sentinel_service_name = mymaster {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600062] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.socket_timeout = 0.1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600230] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.trace_requests = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600390] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler.trace_sqlalchemy = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600572] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler_jaeger.process_tags = {} {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600734] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler_jaeger.service_name_prefix = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.600897] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] profiler_otlp.service_name_prefix = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.601088] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] remote_debug.host = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.601297] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] remote_debug.port = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.601502] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.601672] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.601860] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602052] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602224] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602391] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602552] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602714] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.602881] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603067] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603234] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603407] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603578] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603751] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.603924] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.604108] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.604344] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.604540] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.604711] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.604877] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605059] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605233] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605396] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605565] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605737] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.605927] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.606170] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.606351] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.606526] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.606695] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.606871] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607058] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607229] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607401] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607572] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.ssl_version = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607735] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.607941] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608106] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_notifications.retry = -1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608296] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608477] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_messaging_notifications.transport_url = **** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608648] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.auth_section = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608813] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.auth_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.608976] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.cafile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609153] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.certfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609318] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.collect_timing = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609479] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.connect_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609639] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.connect_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609819] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.endpoint_id = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.609987] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.endpoint_override = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610165] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.insecure = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610324] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.keyfile = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610480] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.max_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610636] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.min_version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610793] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.region_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.610955] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.retriable_status_codes = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611128] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.service_name = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611290] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.service_type = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611451] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.split_loggers = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611609] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.status_code_retries = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611767] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.status_code_retry_delay = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.611956] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.timeout = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612140] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.valid_interfaces = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612302] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_limit.version = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612468] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_reports.file_event_handler = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612631] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612791] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] oslo_reports.log_dir = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.612959] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613136] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613297] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613461] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613627] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613785] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.613957] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614145] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614323] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614494] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614657] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614816] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] vif_plug_ovs_privileged.user = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.614989] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.flat_interface = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.615185] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.615362] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.615535] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.615717] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.615884] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616065] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616233] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616412] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616585] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.isolate_vif = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616753] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.616921] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617103] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617279] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.ovsdb_interface = native {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617442] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_vif_ovs.per_port_bridge = False {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617606] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_brick.lock_path = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617773] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.617947] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618164] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.capabilities = [21] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618330] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618488] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.helper_command = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618653] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618817] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.618976] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] privsep_osbrick.user = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619164] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619327] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.group = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619484] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.helper_command = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619647] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619809] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.619970] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] nova_sys_admin.user = None {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 481.620110] env[62000]: DEBUG oslo_service.service [None req-50f722eb-0559-4d42-add1-b06f52c971af None None] ******************************************************************************** {{(pid=62000) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 481.620610] env[62000]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 482.123805] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Getting list of instances from cluster (obj){ [ 482.123805] env[62000]: value = "domain-c8" [ 482.123805] env[62000]: _type = "ClusterComputeResource" [ 482.123805] env[62000]: } {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 482.125166] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b5ce89-d210-4b1e-b71e-c2a1c169c8fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 482.134376] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Got total of 0 instances {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 482.134908] env[62000]: WARNING nova.virt.vmwareapi.driver [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 482.135380] env[62000]: INFO nova.virt.node [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Generated node identity bcfccf20-49dd-4b91-819e-4373e67bf5ec [ 482.135602] env[62000]: INFO nova.virt.node [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Wrote node identity bcfccf20-49dd-4b91-819e-4373e67bf5ec to /opt/stack/data/n-cpu-1/compute_id [ 482.637925] env[62000]: WARNING nova.compute.manager [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Compute nodes ['bcfccf20-49dd-4b91-819e-4373e67bf5ec'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 483.644581] env[62000]: INFO nova.compute.manager [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 484.649574] env[62000]: WARNING nova.compute.manager [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 484.649956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 484.650134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 484.650292] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 484.650485] env[62000]: DEBUG nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 484.651416] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5781f41-ca41-458c-b1c3-c9932a6f4d42 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.659861] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5f92c5-2e90-4ddd-8c40-f2a92f76a302 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.673379] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62663983-c133-4911-ad79-28136e8db3b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.679980] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d783d7-57f8-4e6d-adb4-8612b24011ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.709883] env[62000]: DEBUG nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181654MB free_disk=53GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 484.710028] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 484.710223] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 485.212794] env[62000]: WARNING nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] No compute node record for cpu-1:bcfccf20-49dd-4b91-819e-4373e67bf5ec: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host bcfccf20-49dd-4b91-819e-4373e67bf5ec could not be found. [ 485.716235] env[62000]: INFO nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: bcfccf20-49dd-4b91-819e-4373e67bf5ec [ 487.226010] env[62000]: DEBUG nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 487.226355] env[62000]: DEBUG nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 487.379350] env[62000]: INFO nova.scheduler.client.report [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] [req-524cd59d-2158-438b-ba90-73219d569e9a] Created resource provider record via placement API for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 487.395754] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2440cbcf-fb19-4199-8012-2518a17b8785 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.403640] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26aa0c29-04d5-4cb0-a540-3be043ddb597 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.432983] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165635b9-8a21-4438-aa54-f7c36272a331 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.439968] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab497ebb-cb18-4cba-b831-064550634310 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.452823] env[62000]: DEBUG nova.compute.provider_tree [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 487.988225] env[62000]: DEBUG nova.scheduler.client.report [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 487.988463] env[62000]: DEBUG nova.compute.provider_tree [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 0 to 1 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 487.988604] env[62000]: DEBUG nova.compute.provider_tree [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 488.037723] env[62000]: DEBUG nova.compute.provider_tree [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 1 to 2 during operation: update_traits {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 488.542405] env[62000]: DEBUG nova.compute.resource_tracker [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 488.542818] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.832s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 488.542818] env[62000]: DEBUG nova.service [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Creating RPC server for service compute {{(pid=62000) start /opt/stack/nova/nova/service.py:186}} [ 488.556458] env[62000]: DEBUG nova.service [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] Join ServiceGroup membership for this service compute {{(pid=62000) start /opt/stack/nova/nova/service.py:203}} [ 488.556633] env[62000]: DEBUG nova.servicegroup.drivers.db [None req-8c7b2b58-11cc-4e9a-ae32-70076fbef255 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62000) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 509.559735] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 510.063361] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Getting list of instances from cluster (obj){ [ 510.063361] env[62000]: value = "domain-c8" [ 510.063361] env[62000]: _type = "ClusterComputeResource" [ 510.063361] env[62000]: } {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 510.064910] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-153b0110-c7bf-40c0-94f5-fb4971f92848 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.073383] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Got total of 0 instances {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 510.073608] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 510.073902] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Getting list of instances from cluster (obj){ [ 510.073902] env[62000]: value = "domain-c8" [ 510.073902] env[62000]: _type = "ClusterComputeResource" [ 510.073902] env[62000]: } {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 510.074744] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a6ba323-d902-4408-b0e1-cb2ee74b8f3e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 510.082975] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Got total of 0 instances {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 524.892888] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquiring lock "e57fb81c-3337-4500-863c-d27eaf763caa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 524.892888] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "e57fb81c-3337-4500-863c-d27eaf763caa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.395419] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 525.796159] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquiring lock "34097eee-1eab-4b71-a450-8d8dddd94276" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.796159] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "34097eee-1eab-4b71-a450-8d8dddd94276" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.952173] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.952742] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.954850] env[62000]: INFO nova.compute.claims [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 526.298405] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.835200] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.860407] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "b8b1038d-86b5-40ce-bc0c-4718013d8e73" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.860723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "b8b1038d-86b5-40ce-bc0c-4718013d8e73" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.036978] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b408d14-bebf-4106-9290-666ac1196fd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.046357] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8148d10-871d-4bc7-b8f7-152fa2a1fd53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.084892] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-904e1b63-6b7c-41a8-8f88-c1882c77c562 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.093585] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48200652-45c6-440c-a28a-e3c4e52ec634 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.110589] env[62000]: DEBUG nova.compute.provider_tree [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 527.363560] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 527.614411] env[62000]: DEBUG nova.scheduler.client.report [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.892375] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.059109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquiring lock "4489993f-7d78-4891-8282-5d06d7689447" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.060112] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "4489993f-7d78-4891-8282-5d06d7689447" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.119501] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.167s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.119736] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 528.125422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.291s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.127995] env[62000]: INFO nova.compute.claims [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 528.562202] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.627082] env[62000]: DEBUG nova.compute.utils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 528.628788] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 528.630156] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 529.091938] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.138760] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 529.251387] env[62000]: DEBUG nova.policy [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a18bfe5c0864f02989140e36f12a5d3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c551e1575bb94ff7914bc46f3be5ba18', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 529.266555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9930bc33-e9be-4411-92c0-b8cd54152fdf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.275129] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3bf6a0-23fc-467b-90a4-fedd163ae626 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.309219] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f318d3b5-5f95-412e-a9db-6a83454dcda7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.316682] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a765a857-ee6f-428b-9ba6-9f3b06053b95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.330604] env[62000]: DEBUG nova.compute.provider_tree [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 529.552960] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquiring lock "6a67dde1-ca25-493e-8d65-8d4cea487c85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.552960] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "6a67dde1-ca25-493e-8d65-8d4cea487c85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.833453] env[62000]: DEBUG nova.scheduler.client.report [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 530.053156] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 530.149343] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 530.193761] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 530.194500] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 530.194500] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 530.194500] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 530.194640] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 530.194674] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 530.194879] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 530.195038] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 530.195543] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 530.195619] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 530.195771] env[62000]: DEBUG nova.virt.hardware [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 530.196685] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34f937a8-a9fb-4935-b97b-99930cf661e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.208015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f5aa07-54d6-4f97-8078-1577487bb73d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.226617] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c645086-4233-4bbd-a8b7-9422bae9e10b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.344616] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.219s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 530.346134] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 530.350865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.458s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.352460] env[62000]: INFO nova.compute.claims [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.446863] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Successfully created port: 93c52043-f8ca-4ddc-a31e-a33688e9e6ab {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 530.595403] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.853243] env[62000]: DEBUG nova.compute.utils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.855243] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.855361] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.999802] env[62000]: DEBUG nova.policy [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '26cfebf7f10047fbb4146a102acf510d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2fd94254cf5c4d46a611c4dcfd2e5abf', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 531.360119] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 531.473448] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-773a565b-076d-40b8-8590-f7c6bce9351f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.483417] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653bf33f-78b6-48ef-87c2-b68bc1df6b2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.515775] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cb52b4-3928-45cc-9219-cac127e8bcfe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.526307] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d220a58-5661-4a25-b88d-3eaf2d516524 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.542112] env[62000]: DEBUG nova.compute.provider_tree [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.865517] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Successfully created port: 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.046096] env[62000]: DEBUG nova.scheduler.client.report [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.375105] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 532.412545] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 532.412545] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 532.412545] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 532.412960] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 532.412960] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 532.412960] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 532.412960] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 532.412960] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 532.413130] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 532.413968] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 532.414207] env[62000]: DEBUG nova.virt.hardware [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 532.415149] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a2ebe6c-380c-4793-87b4-d953c08c8522 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.426600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa347a8f-5aeb-4f25-b828-c5eeb6b3829a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.555422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.555923] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 532.560361] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.469s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.561801] env[62000]: INFO nova.compute.claims [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.874311] env[62000]: ERROR nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 532.874311] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.874311] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.874311] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.874311] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.874311] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.874311] env[62000]: ERROR nova.compute.manager raise self.value [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.874311] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.874311] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.874311] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.874897] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.874897] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.874897] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 532.874897] env[62000]: ERROR nova.compute.manager [ 532.874897] env[62000]: Traceback (most recent call last): [ 532.874897] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.874897] env[62000]: listener.cb(fileno) [ 532.874897] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.874897] env[62000]: result = function(*args, **kwargs) [ 532.874897] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.874897] env[62000]: return func(*args, **kwargs) [ 532.874897] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.874897] env[62000]: raise e [ 532.874897] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.874897] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 532.874897] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.874897] env[62000]: created_port_ids = self._update_ports_for_instance( [ 532.874897] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.874897] env[62000]: with excutils.save_and_reraise_exception(): [ 532.874897] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.874897] env[62000]: self.force_reraise() [ 532.874897] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.874897] env[62000]: raise self.value [ 532.874897] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.874897] env[62000]: updated_port = self._update_port( [ 532.874897] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.874897] env[62000]: _ensure_no_port_binding_failure(port) [ 532.874897] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.874897] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.878070] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 532.878070] env[62000]: Removing descriptor: 14 [ 532.878070] env[62000]: ERROR nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Traceback (most recent call last): [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] yield resources [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.driver.spawn(context, instance, image_meta, [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.878070] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] vm_ref = self.build_virtual_machine(instance, [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] for vif in network_info: [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self._sync_wrapper(fn, *args, **kwargs) [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.wait() [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self[:] = self._gt.wait() [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self._exit_event.wait() [ 532.878377] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] result = hub.switch() [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self.greenlet.switch() [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] result = function(*args, **kwargs) [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return func(*args, **kwargs) [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise e [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] nwinfo = self.network_api.allocate_for_instance( [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 532.878678] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] created_port_ids = self._update_ports_for_instance( [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] with excutils.save_and_reraise_exception(): [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.force_reraise() [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise self.value [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] updated_port = self._update_port( [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] _ensure_no_port_binding_failure(port) [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.878976] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise exception.PortBindingFailed(port_id=port['id']) [ 532.879311] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 532.879311] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] [ 532.879311] env[62000]: INFO nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Terminating instance [ 532.881107] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquiring lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.881722] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquired lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.882174] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.066110] env[62000]: DEBUG nova.compute.utils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 533.066110] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 533.066110] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 533.168076] env[62000]: DEBUG nova.policy [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9ab8430f2275456a9b16bb2c3573e6f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a070b994c52040839c485058d9500eb9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 533.419926] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 533.573809] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 533.642348] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "dc980828-985b-4da1-b3fe-e3cda8b84a6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.642348] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "dc980828-985b-4da1-b3fe-e3cda8b84a6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.690113] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "e99bfd97-1952-4166-aa19-560053bae307" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.690113] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.723786] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "fa76f279-9efd-401f-823e-806425ce39e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.724089] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.777555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ad7d46-c039-406d-be36-4d3e5268515f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.787890] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5d2ca33-8985-4e75-8429-32b45e2f5473 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.823670] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3cec831-4d20-43c3-9669-48c484895942 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.827617] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.834533] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1834c9-528e-4b94-8b44-1e3312835ea7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.852603] env[62000]: DEBUG nova.compute.provider_tree [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.153485] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.189337] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Successfully created port: 220d283b-60ac-4137-aeaf-43e249ee516f {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.191544] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.227664] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 534.330819] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Releasing lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.331881] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 534.331881] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 534.332214] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dfec47c-0b02-4c19-ab45-2cefd963cb4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.344484] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2b97c00-7ded-4ff6-a127-f2273b505986 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.356368] env[62000]: DEBUG nova.scheduler.client.report [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.371719] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e57fb81c-3337-4500-863c-d27eaf763caa could not be found. [ 534.372107] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.372471] env[62000]: INFO nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Took 0.04 seconds to destroy the instance on the hypervisor. [ 534.372751] env[62000]: DEBUG oslo.service.loopingcall [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.373063] env[62000]: DEBUG nova.compute.manager [-] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.373163] env[62000]: DEBUG nova.network.neutron [-] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 534.403410] env[62000]: DEBUG nova.network.neutron [-] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.595196] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 534.626770] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.626952] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.627108] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.627293] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.627437] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.627591] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.627775] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.628023] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.628108] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.628270] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.628440] env[62000]: DEBUG nova.virt.hardware [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.629804] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326d97df-b8a7-4f89-b1b2-d7f36eb2af47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.638656] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dbb7f5-4138-4867-aaa7-2a7d7262cbda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.693579] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.716686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.755424] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.863349] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.863349] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.869279] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.273s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.869783] env[62000]: INFO nova.compute.claims [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.908249] env[62000]: DEBUG nova.network.neutron [-] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.104019] env[62000]: DEBUG nova.compute.manager [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Received event network-changed-93c52043-f8ca-4ddc-a31e-a33688e9e6ab {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 535.104019] env[62000]: DEBUG nova.compute.manager [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Refreshing instance network info cache due to event network-changed-93c52043-f8ca-4ddc-a31e-a33688e9e6ab. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 535.104019] env[62000]: DEBUG oslo_concurrency.lockutils [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] Acquiring lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.104019] env[62000]: DEBUG oslo_concurrency.lockutils [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] Acquired lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.105637] env[62000]: DEBUG nova.network.neutron [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Refreshing network info cache for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.378692] env[62000]: DEBUG nova.compute.utils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.380100] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 535.380419] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 535.410710] env[62000]: INFO nova.compute.manager [-] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Took 1.04 seconds to deallocate network for instance. [ 535.413020] env[62000]: DEBUG nova.compute.claims [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 535.413141] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.444648] env[62000]: ERROR nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 535.444648] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.444648] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.444648] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.444648] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.444648] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.444648] env[62000]: ERROR nova.compute.manager raise self.value [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.444648] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 535.444648] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.444648] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 535.445185] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.445185] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 535.445185] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 535.445185] env[62000]: ERROR nova.compute.manager [ 535.445185] env[62000]: Traceback (most recent call last): [ 535.445185] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 535.445185] env[62000]: listener.cb(fileno) [ 535.445185] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.445185] env[62000]: result = function(*args, **kwargs) [ 535.445185] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.445185] env[62000]: return func(*args, **kwargs) [ 535.445185] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.445185] env[62000]: raise e [ 535.445185] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.445185] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 535.445185] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.445185] env[62000]: created_port_ids = self._update_ports_for_instance( [ 535.445185] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.445185] env[62000]: with excutils.save_and_reraise_exception(): [ 535.445185] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.445185] env[62000]: self.force_reraise() [ 535.445185] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.445185] env[62000]: raise self.value [ 535.445185] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.445185] env[62000]: updated_port = self._update_port( [ 535.445185] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.445185] env[62000]: _ensure_no_port_binding_failure(port) [ 535.445185] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.445185] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 535.445877] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 535.445877] env[62000]: Removing descriptor: 15 [ 535.445877] env[62000]: ERROR nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Traceback (most recent call last): [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] yield resources [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.driver.spawn(context, instance, image_meta, [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self._vmops.spawn(context, instance, image_meta, injected_files, [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 535.445877] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] vm_ref = self.build_virtual_machine(instance, [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] vif_infos = vmwarevif.get_vif_info(self._session, [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] for vif in network_info: [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self._sync_wrapper(fn, *args, **kwargs) [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.wait() [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self[:] = self._gt.wait() [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self._exit_event.wait() [ 535.446275] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] result = hub.switch() [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self.greenlet.switch() [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] result = function(*args, **kwargs) [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return func(*args, **kwargs) [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise e [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] nwinfo = self.network_api.allocate_for_instance( [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 535.446600] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] created_port_ids = self._update_ports_for_instance( [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] with excutils.save_and_reraise_exception(): [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.force_reraise() [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise self.value [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] updated_port = self._update_port( [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] _ensure_no_port_binding_failure(port) [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 535.446906] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise exception.PortBindingFailed(port_id=port['id']) [ 535.447192] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 535.447192] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] [ 535.447192] env[62000]: INFO nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Terminating instance [ 535.450055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquiring lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 535.450055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquired lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.450055] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 535.501368] env[62000]: DEBUG nova.policy [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b2d70b1add274c49abbba2e252c6fe78', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '907227f8debb4a1a97cf478a32e57119', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 535.642794] env[62000]: DEBUG nova.network.neutron [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.727088] env[62000]: DEBUG nova.network.neutron [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.884938] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.002319] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.057758] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e92bd245-c760-413d-8b1c-444e05d35f88 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.067569] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4283e09-ca28-42ae-95c5-bb03d7f78dcc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.109228] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d067ab9-74f4-46dc-bd9d-3bdaf2e33083 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.113940] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.118986] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c6f3409-fd76-41c7-ac0a-7988af1a7f40 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.135406] env[62000]: DEBUG nova.compute.provider_tree [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.152865] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Successfully created port: 15d1f9f0-7e00-4389-b925-9af9d7b68ef9 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.231662] env[62000]: DEBUG oslo_concurrency.lockutils [req-d6219b86-ed73-489a-8f89-f0bdc5a50d24 req-dd52d261-91c3-4570-b9f0-2510baf3ef20 service nova] Releasing lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.312800] env[62000]: DEBUG nova.compute.manager [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Received event network-changed-70f47b70-cc2d-4cb1-a9e4-63bd02b4407e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 536.313906] env[62000]: DEBUG nova.compute.manager [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Refreshing instance network info cache due to event network-changed-70f47b70-cc2d-4cb1-a9e4-63bd02b4407e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 536.313906] env[62000]: DEBUG oslo_concurrency.lockutils [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] Acquiring lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.523087] env[62000]: ERROR nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 536.523087] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.523087] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.523087] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.523087] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.523087] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.523087] env[62000]: ERROR nova.compute.manager raise self.value [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.523087] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.523087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.523087] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.523886] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.523886] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.523886] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 536.523886] env[62000]: ERROR nova.compute.manager [ 536.523886] env[62000]: Traceback (most recent call last): [ 536.523886] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.523886] env[62000]: listener.cb(fileno) [ 536.523886] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.523886] env[62000]: result = function(*args, **kwargs) [ 536.523886] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.523886] env[62000]: return func(*args, **kwargs) [ 536.523886] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.523886] env[62000]: raise e [ 536.523886] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.523886] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 536.523886] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.523886] env[62000]: created_port_ids = self._update_ports_for_instance( [ 536.523886] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.523886] env[62000]: with excutils.save_and_reraise_exception(): [ 536.523886] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.523886] env[62000]: self.force_reraise() [ 536.523886] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.523886] env[62000]: raise self.value [ 536.523886] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.523886] env[62000]: updated_port = self._update_port( [ 536.523886] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.523886] env[62000]: _ensure_no_port_binding_failure(port) [ 536.523886] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.523886] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.524666] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 536.524666] env[62000]: Removing descriptor: 14 [ 536.524666] env[62000]: ERROR nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Traceback (most recent call last): [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] yield resources [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.driver.spawn(context, instance, image_meta, [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.524666] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] vm_ref = self.build_virtual_machine(instance, [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] for vif in network_info: [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self._sync_wrapper(fn, *args, **kwargs) [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.wait() [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self[:] = self._gt.wait() [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self._exit_event.wait() [ 536.525019] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] result = hub.switch() [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self.greenlet.switch() [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] result = function(*args, **kwargs) [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return func(*args, **kwargs) [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise e [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] nwinfo = self.network_api.allocate_for_instance( [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.525405] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] created_port_ids = self._update_ports_for_instance( [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] with excutils.save_and_reraise_exception(): [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.force_reraise() [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise self.value [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] updated_port = self._update_port( [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] _ensure_no_port_binding_failure(port) [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.525761] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise exception.PortBindingFailed(port_id=port['id']) [ 536.526118] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 536.526118] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] [ 536.526118] env[62000]: INFO nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Terminating instance [ 536.529115] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.530071] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquired lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.530176] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.619649] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Releasing lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.620133] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 536.620343] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 536.620647] env[62000]: DEBUG oslo_concurrency.lockutils [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] Acquired lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.620822] env[62000]: DEBUG nova.network.neutron [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Refreshing network info cache for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 536.621897] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fb618fe4-be92-4ea2-86ea-dc59e94c2395 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.635232] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b44a17-890b-4ac5-b218-e809d7c32e88 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.648902] env[62000]: DEBUG nova.scheduler.client.report [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.665313] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 34097eee-1eab-4b71-a450-8d8dddd94276 could not be found. [ 536.665313] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 536.665313] env[62000]: INFO nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Took 0.04 seconds to destroy the instance on the hypervisor. [ 536.665313] env[62000]: DEBUG oslo.service.loopingcall [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.665449] env[62000]: DEBUG nova.compute.manager [-] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 536.665449] env[62000]: DEBUG nova.network.neutron [-] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 536.712040] env[62000]: DEBUG nova.network.neutron [-] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.825338] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "15409fa7-254c-435e-8080-46d3f65b2d46" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.826294] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.897720] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.934091] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.934091] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.935963] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.936342] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.936465] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.936648] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.936883] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.937092] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.937320] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.937523] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.937748] env[62000]: DEBUG nova.virt.hardware [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.938672] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10b0fffc-7151-4594-a8e5-976890925a39 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.950817] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d07cdc-900f-4f07-9d4f-f93c09cd4f87 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.039512] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.039746] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.042835] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 537.042835] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Rebuilding the list of instances to heal {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 537.149120] env[62000]: DEBUG nova.network.neutron [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.153815] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.285s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.154457] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 537.160864] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.466s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.165019] env[62000]: INFO nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.178779] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.215575] env[62000]: DEBUG nova.network.neutron [-] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.245026] env[62000]: DEBUG nova.network.neutron [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.329123] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.366377] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.545797] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 537.545961] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 537.546115] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 537.546241] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 537.546362] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 537.546482] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Didn't find any instances for network info cache update. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 537.546948] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.548106] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.548324] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.548532] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.548703] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.548883] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.549062] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 537.549225] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 537.670869] env[62000]: DEBUG nova.compute.utils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.677608] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 537.677608] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.721898] env[62000]: INFO nova.compute.manager [-] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Took 1.05 seconds to deallocate network for instance. [ 537.724095] env[62000]: DEBUG nova.compute.claims [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 537.724095] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.750259] env[62000]: DEBUG oslo_concurrency.lockutils [req-4c310449-5d5a-4f6a-a8d1-34d76e8713f2 req-e7e25e79-2c6e-4dab-be80-c03e495a3ac3 service nova] Releasing lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.833643] env[62000]: DEBUG nova.policy [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a744ab2e149d401cb723f5ff8037d426', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5805a97880224a5fa6feb5471ded28a0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 537.862466] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.868653] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Releasing lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.869089] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 537.869357] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 537.869547] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c178910d-9127-4c44-b6ea-73720234d98d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.882683] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2833e4a-4746-4595-9f34-0543d49d3e6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.906478] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b8b1038d-86b5-40ce-bc0c-4718013d8e73 could not be found. [ 537.906739] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 537.906918] env[62000]: INFO nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Took 0.04 seconds to destroy the instance on the hypervisor. [ 537.907841] env[62000]: DEBUG oslo.service.loopingcall [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.907841] env[62000]: DEBUG nova.compute.manager [-] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 537.907841] env[62000]: DEBUG nova.network.neutron [-] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.975127] env[62000]: DEBUG nova.network.neutron [-] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.052422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.175836] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 538.379140] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81800c28-c2df-4d15-b747-fae1ce90d75d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.388047] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93e76281-c6ef-4e4a-bb5b-829230677fc1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.425950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39908138-a786-44f4-bdb9-347f0e17bae6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.436245] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21629cfe-2f7f-4f31-bdf4-d76a2f2bad7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.451972] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.458758] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquiring lock "faa1794d-5941-40fb-bd78-2608fa2d5da0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.458758] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "faa1794d-5941-40fb-bd78-2608fa2d5da0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.478857] env[62000]: DEBUG nova.network.neutron [-] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.737251] env[62000]: DEBUG nova.compute.manager [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Received event network-vif-deleted-93c52043-f8ca-4ddc-a31e-a33688e9e6ab {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 538.737606] env[62000]: DEBUG nova.compute.manager [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Received event network-changed-220d283b-60ac-4137-aeaf-43e249ee516f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 538.737606] env[62000]: DEBUG nova.compute.manager [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Refreshing instance network info cache due to event network-changed-220d283b-60ac-4137-aeaf-43e249ee516f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 538.740574] env[62000]: DEBUG oslo_concurrency.lockutils [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] Acquiring lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.740788] env[62000]: DEBUG oslo_concurrency.lockutils [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] Acquired lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.740892] env[62000]: DEBUG nova.network.neutron [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Refreshing network info cache for port 220d283b-60ac-4137-aeaf-43e249ee516f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 538.878891] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Successfully created port: 9fc62892-3187-47ea-b8d5-13a99d49ddfb {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.958147] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.963735] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 538.983142] env[62000]: INFO nova.compute.manager [-] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Took 1.08 seconds to deallocate network for instance. [ 538.986407] env[62000]: DEBUG nova.compute.claims [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 538.986718] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.198194] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 539.239484] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 539.239770] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 539.239885] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 539.240547] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 539.240547] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 539.240547] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 539.240839] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 539.240839] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 539.240950] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 539.241248] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 539.242227] env[62000]: DEBUG nova.virt.hardware [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 539.242383] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43438d89-31a1-4b99-b5dc-f082fd31a349 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.253178] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bfd4d0a-6091-42a6-bd17-84cd1956efd8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.306201] env[62000]: DEBUG nova.compute.manager [req-ac0d55e5-e5ef-4b54-a015-da5a09a68740 req-db19cf88-d125-402c-85d3-c4b9b9631cc6 service nova] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Received event network-vif-deleted-70f47b70-cc2d-4cb1-a9e4-63bd02b4407e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 539.322445] env[62000]: DEBUG nova.network.neutron [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.473592] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.474124] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.480457] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.764s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.482250] env[62000]: INFO nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 539.505237] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.638880] env[62000]: DEBUG nova.network.neutron [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.981721] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.983549] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 539.983767] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 540.051180] env[62000]: ERROR nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 540.051180] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.051180] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.051180] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.051180] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.051180] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.051180] env[62000]: ERROR nova.compute.manager raise self.value [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.051180] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.051180] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.051180] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.051588] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.051588] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.051588] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 540.051588] env[62000]: ERROR nova.compute.manager [ 540.051588] env[62000]: Traceback (most recent call last): [ 540.051588] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.051588] env[62000]: listener.cb(fileno) [ 540.051588] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.051588] env[62000]: result = function(*args, **kwargs) [ 540.051588] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.051588] env[62000]: return func(*args, **kwargs) [ 540.051588] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.051588] env[62000]: raise e [ 540.051588] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.051588] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 540.051588] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.051588] env[62000]: created_port_ids = self._update_ports_for_instance( [ 540.051588] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.051588] env[62000]: with excutils.save_and_reraise_exception(): [ 540.051588] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.051588] env[62000]: self.force_reraise() [ 540.051588] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.051588] env[62000]: raise self.value [ 540.051588] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.051588] env[62000]: updated_port = self._update_port( [ 540.051588] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.051588] env[62000]: _ensure_no_port_binding_failure(port) [ 540.051588] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.051588] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.052359] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 540.052359] env[62000]: Removing descriptor: 16 [ 540.052359] env[62000]: ERROR nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] Traceback (most recent call last): [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] yield resources [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.driver.spawn(context, instance, image_meta, [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.052359] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] vm_ref = self.build_virtual_machine(instance, [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] for vif in network_info: [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self._sync_wrapper(fn, *args, **kwargs) [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.wait() [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self[:] = self._gt.wait() [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self._exit_event.wait() [ 540.052675] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] result = hub.switch() [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self.greenlet.switch() [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] result = function(*args, **kwargs) [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return func(*args, **kwargs) [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise e [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] nwinfo = self.network_api.allocate_for_instance( [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.053038] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] created_port_ids = self._update_ports_for_instance( [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] with excutils.save_and_reraise_exception(): [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.force_reraise() [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise self.value [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] updated_port = self._update_port( [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] _ensure_no_port_binding_failure(port) [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.053349] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise exception.PortBindingFailed(port_id=port['id']) [ 540.053702] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 540.053702] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] [ 540.053702] env[62000]: INFO nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Terminating instance [ 540.054578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquiring lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.054712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquired lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.054911] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.125933] env[62000]: DEBUG nova.policy [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fe0c77929604f70b2f8c4c2d11bb017', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6a3e788e8c6491ca52b042247c5b9f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.142259] env[62000]: DEBUG oslo_concurrency.lockutils [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] Releasing lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.142617] env[62000]: DEBUG nova.compute.manager [req-18e8e2a0-f266-4ad0-8e71-d66f3ebeaf44 req-bea5b2f6-ad5b-42bc-b1ca-bb6a7f4edff2 service nova] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Received event network-vif-deleted-220d283b-60ac-4137-aeaf-43e249ee516f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 540.489303] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.621722] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.669025] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650af391-38ac-45f4-80fb-33672016a221 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.677884] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c00a055-2f66-4d08-a388-bf7973c3c873 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.714631] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13cf1e13-82da-4ed8-972e-7bfc14ef61d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.722813] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e712d2d4-d028-4bc8-93ff-cbdc2931173b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.739162] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.070566] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.243051] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.464278] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Successfully created port: 39817f72-1f81-4f77-8c09-d9cf7fde1df4 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.501376] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 541.543039] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 541.543039] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 541.543039] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 541.543254] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 541.543254] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 541.543254] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 541.543254] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 541.543899] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 541.544229] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 541.544592] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 541.546068] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 541.547016] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3615bfc5-2f4e-4470-8254-30bdb13e3054 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.557940] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7911920e-b013-43d5-a3d0-0cf2473274d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.578017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Releasing lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.578732] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.579074] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 541.579764] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbb17f2b-5d90-4485-af8a-51f135eb2f84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.592039] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44702877-321a-48c3-9e29-86a835c19737 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.616497] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4489993f-7d78-4891-8282-5d06d7689447 could not be found. [ 541.616717] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 541.616907] env[62000]: INFO nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Took 0.04 seconds to destroy the instance on the hypervisor. [ 541.617168] env[62000]: DEBUG oslo.service.loopingcall [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.617383] env[62000]: DEBUG nova.compute.manager [-] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.617479] env[62000]: DEBUG nova.network.neutron [-] [instance: 4489993f-7d78-4891-8282-5d06d7689447] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.676403] env[62000]: DEBUG nova.network.neutron [-] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.748055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.268s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.748598] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 541.751215] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.996s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.752610] env[62000]: INFO nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.179450] env[62000]: DEBUG nova.network.neutron [-] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.260936] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 542.263184] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 542.263415] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 542.334548] env[62000]: DEBUG nova.compute.manager [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Received event network-changed-15d1f9f0-7e00-4389-b925-9af9d7b68ef9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.334548] env[62000]: DEBUG nova.compute.manager [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Refreshing instance network info cache due to event network-changed-15d1f9f0-7e00-4389-b925-9af9d7b68ef9. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 542.334548] env[62000]: DEBUG oslo_concurrency.lockutils [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] Acquiring lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.334548] env[62000]: DEBUG oslo_concurrency.lockutils [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] Acquired lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.334548] env[62000]: DEBUG nova.network.neutron [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Refreshing network info cache for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 542.397069] env[62000]: DEBUG nova.policy [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fe0c77929604f70b2f8c4c2d11bb017', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6a3e788e8c6491ca52b042247c5b9f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 542.682325] env[62000]: INFO nova.compute.manager [-] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Took 1.06 seconds to deallocate network for instance. [ 542.687965] env[62000]: DEBUG nova.compute.claims [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 542.688187] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.767162] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 542.952902] env[62000]: DEBUG nova.network.neutron [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.984548] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc638bb9-8337-473f-b6c1-39e1a8251a38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.994969] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50502c04-7550-4a78-a309-0713f057582b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.032173] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131b936e-35b5-44ab-ac85-cac15ae56e1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.038156] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93edb12-2fb3-4bf8-9b26-eff8bbe77a45 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.052508] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.418354] env[62000]: DEBUG nova.network.neutron [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.556726] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.600477] env[62000]: ERROR nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 543.600477] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.600477] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.600477] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.600477] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.600477] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.600477] env[62000]: ERROR nova.compute.manager raise self.value [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.600477] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.600477] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.600477] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.600961] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.600961] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.600961] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 543.600961] env[62000]: ERROR nova.compute.manager [ 543.600961] env[62000]: Traceback (most recent call last): [ 543.600961] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.600961] env[62000]: listener.cb(fileno) [ 543.600961] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.600961] env[62000]: result = function(*args, **kwargs) [ 543.600961] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.600961] env[62000]: return func(*args, **kwargs) [ 543.600961] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.600961] env[62000]: raise e [ 543.600961] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.600961] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 543.600961] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.600961] env[62000]: created_port_ids = self._update_ports_for_instance( [ 543.600961] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.600961] env[62000]: with excutils.save_and_reraise_exception(): [ 543.600961] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.600961] env[62000]: self.force_reraise() [ 543.600961] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.600961] env[62000]: raise self.value [ 543.600961] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.600961] env[62000]: updated_port = self._update_port( [ 543.600961] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.600961] env[62000]: _ensure_no_port_binding_failure(port) [ 543.600961] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.600961] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.601784] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 543.601784] env[62000]: Removing descriptor: 14 [ 543.601784] env[62000]: ERROR nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Traceback (most recent call last): [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] yield resources [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.driver.spawn(context, instance, image_meta, [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.601784] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] vm_ref = self.build_virtual_machine(instance, [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] for vif in network_info: [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self._sync_wrapper(fn, *args, **kwargs) [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.wait() [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self[:] = self._gt.wait() [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self._exit_event.wait() [ 543.602133] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] result = hub.switch() [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self.greenlet.switch() [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] result = function(*args, **kwargs) [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return func(*args, **kwargs) [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise e [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] nwinfo = self.network_api.allocate_for_instance( [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.602465] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] created_port_ids = self._update_ports_for_instance( [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] with excutils.save_and_reraise_exception(): [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.force_reraise() [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise self.value [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] updated_port = self._update_port( [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] _ensure_no_port_binding_failure(port) [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.602840] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise exception.PortBindingFailed(port_id=port['id']) [ 543.603282] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 543.603282] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] [ 543.603282] env[62000]: INFO nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Terminating instance [ 543.608362] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquiring lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.608362] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquired lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.608362] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 543.785900] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 543.816252] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.816252] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.816252] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.820018] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.820018] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.820018] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.820018] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.820018] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.820331] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.820331] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.820331] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.820331] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f9b1ec-8c76-45b6-80e3-bde31daa7310 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.824110] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Successfully created port: 5fe122d3-6f5e-40c9-b7e5-31fd07e29683 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 543.834601] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf8d6e5-52ef-4a49-887c-a550c728abcc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.925194] env[62000]: DEBUG oslo_concurrency.lockutils [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] Releasing lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.925194] env[62000]: DEBUG nova.compute.manager [req-3dd6f86f-3042-488f-9642-cc2e89dae7dc req-445fc1ab-2854-48c1-9563-653071a58829 service nova] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Received event network-vif-deleted-15d1f9f0-7e00-4389-b925-9af9d7b68ef9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 544.068182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.317s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.069210] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 544.075588] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.662s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.182305] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 544.492965] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.585198] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.593717] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 544.593885] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 544.746544] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquiring lock "57cc0a22-28ff-4a24-9fea-5c6a0a699a67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.747148] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "57cc0a22-28ff-4a24-9fea-5c6a0a699a67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.770938] env[62000]: DEBUG nova.policy [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8fe0c77929604f70b2f8c4c2d11bb017', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a6a3e788e8c6491ca52b042247c5b9f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 544.776792] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-646918c7-c4b9-447e-be56-af867680d4a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.784753] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d11b2259-e25f-4980-bc25-c1a1a8326514 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.814408] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb5d739-9860-4e22-9340-46d6c86d87ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.822554] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67871b8f-62e2-4a0a-87c9-24a10d5e610e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.837418] env[62000]: DEBUG nova.compute.provider_tree [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 544.997493] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Releasing lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.998039] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.998147] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 544.998418] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b02cd207-9f5b-4089-9665-c3e7c4e4c175 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.010256] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9cba922-4ca9-4639-b152-3206191c0a28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.034869] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a67dde1-ca25-493e-8d65-8d4cea487c85 could not be found. [ 545.035197] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 545.035324] env[62000]: INFO nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Took 0.04 seconds to destroy the instance on the hypervisor. [ 545.035552] env[62000]: DEBUG oslo.service.loopingcall [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 545.035977] env[62000]: DEBUG nova.compute.manager [-] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 545.035977] env[62000]: DEBUG nova.network.neutron [-] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.094492] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.127549] env[62000]: DEBUG nova.network.neutron [-] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.344361] env[62000]: DEBUG nova.scheduler.client.report [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 545.632405] env[62000]: DEBUG nova.network.neutron [-] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.851096] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.775s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.852257] env[62000]: ERROR nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Traceback (most recent call last): [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.driver.spawn(context, instance, image_meta, [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self._vmops.spawn(context, instance, image_meta, injected_files, [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] vm_ref = self.build_virtual_machine(instance, [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] vif_infos = vmwarevif.get_vif_info(self._session, [ 545.852257] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] for vif in network_info: [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self._sync_wrapper(fn, *args, **kwargs) [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.wait() [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self[:] = self._gt.wait() [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self._exit_event.wait() [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] result = hub.switch() [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 545.852610] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return self.greenlet.switch() [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] result = function(*args, **kwargs) [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] return func(*args, **kwargs) [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise e [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] nwinfo = self.network_api.allocate_for_instance( [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] created_port_ids = self._update_ports_for_instance( [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] with excutils.save_and_reraise_exception(): [ 545.853090] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] self.force_reraise() [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise self.value [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] updated_port = self._update_port( [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] _ensure_no_port_binding_failure(port) [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] raise exception.PortBindingFailed(port_id=port['id']) [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] nova.exception.PortBindingFailed: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. [ 545.853415] env[62000]: ERROR nova.compute.manager [instance: e57fb81c-3337-4500-863c-d27eaf763caa] [ 545.854197] env[62000]: DEBUG nova.compute.utils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 545.856661] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.133s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 545.864605] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Build of instance e57fb81c-3337-4500-863c-d27eaf763caa was re-scheduled: Binding failed for port 93c52043-f8ca-4ddc-a31e-a33688e9e6ab, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 545.865085] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 545.865362] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquiring lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.865516] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Acquired lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.865703] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.105192] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 546.127172] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.127445] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.127599] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.127775] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.127917] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.128077] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.128289] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.128442] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.128602] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.128758] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.128928] env[62000]: DEBUG nova.virt.hardware [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.130126] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8977282-2d04-4f1f-91f1-5a42ec4442c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.134710] env[62000]: INFO nova.compute.manager [-] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Took 1.10 seconds to deallocate network for instance. [ 546.139268] env[62000]: DEBUG nova.compute.claims [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 546.139438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.141059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f98c242d-e41b-4e6b-bde6-2f08372a0fa6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.357407] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Successfully created port: fa404718-9b22-4a2f-a565-58fd495c93fe {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 546.449802] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.548187] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1442111-79b6-4745-8b80-87a7ba1faaea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.557379] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72effcf-c87b-4d16-b65d-bd18bd37616d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.595413] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9d5d8b-a512-4ff6-af7a-9450c03b0ffc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.604627] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a02bbd7-fbfd-4f04-9732-0acadbf24c0a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.623756] env[62000]: DEBUG nova.compute.provider_tree [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.935452] env[62000]: DEBUG nova.compute.manager [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Received event network-changed-9fc62892-3187-47ea-b8d5-13a99d49ddfb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 546.935692] env[62000]: DEBUG nova.compute.manager [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Refreshing instance network info cache due to event network-changed-9fc62892-3187-47ea-b8d5-13a99d49ddfb. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 546.935729] env[62000]: DEBUG oslo_concurrency.lockutils [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] Acquiring lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.935865] env[62000]: DEBUG oslo_concurrency.lockutils [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] Acquired lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.936908] env[62000]: DEBUG nova.network.neutron [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Refreshing network info cache for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 546.957723] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.084909] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 547.084909] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.084909] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.084909] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.084909] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.084909] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.084909] env[62000]: ERROR nova.compute.manager raise self.value [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.084909] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.084909] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.084909] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.085559] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.085559] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.085559] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 547.085559] env[62000]: ERROR nova.compute.manager [ 547.085559] env[62000]: Traceback (most recent call last): [ 547.085559] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.085559] env[62000]: listener.cb(fileno) [ 547.085559] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.085559] env[62000]: result = function(*args, **kwargs) [ 547.085559] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.085559] env[62000]: return func(*args, **kwargs) [ 547.085559] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.085559] env[62000]: raise e [ 547.085559] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.085559] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 547.085559] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.085559] env[62000]: created_port_ids = self._update_ports_for_instance( [ 547.085559] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.085559] env[62000]: with excutils.save_and_reraise_exception(): [ 547.085559] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.085559] env[62000]: self.force_reraise() [ 547.085559] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.085559] env[62000]: raise self.value [ 547.085559] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.085559] env[62000]: updated_port = self._update_port( [ 547.085559] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.085559] env[62000]: _ensure_no_port_binding_failure(port) [ 547.085559] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.085559] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.086669] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 547.086669] env[62000]: Removing descriptor: 15 [ 547.086669] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Traceback (most recent call last): [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] yield resources [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.driver.spawn(context, instance, image_meta, [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.086669] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] vm_ref = self.build_virtual_machine(instance, [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] for vif in network_info: [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self._sync_wrapper(fn, *args, **kwargs) [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.wait() [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self[:] = self._gt.wait() [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self._exit_event.wait() [ 547.087028] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] result = hub.switch() [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self.greenlet.switch() [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] result = function(*args, **kwargs) [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return func(*args, **kwargs) [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise e [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] nwinfo = self.network_api.allocate_for_instance( [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.087477] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] created_port_ids = self._update_ports_for_instance( [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] with excutils.save_and_reraise_exception(): [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.force_reraise() [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise self.value [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] updated_port = self._update_port( [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] _ensure_no_port_binding_failure(port) [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.087883] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise exception.PortBindingFailed(port_id=port['id']) [ 547.088363] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 547.088363] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] [ 547.088363] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Terminating instance [ 547.089110] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.089207] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.089831] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.127211] env[62000]: DEBUG nova.scheduler.client.report [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 547.237345] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquiring lock "c0f34d3e-39a1-4487-bd1f-ade41ef6832c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.237572] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "c0f34d3e-39a1-4487-bd1f-ade41ef6832c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.461414] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Releasing lock "refresh_cache-e57fb81c-3337-4500-863c-d27eaf763caa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.461414] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 547.461414] env[62000]: DEBUG nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 547.461414] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 547.497639] env[62000]: DEBUG nova.network.neutron [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.537057] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.633420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.777s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.634220] env[62000]: ERROR nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Traceback (most recent call last): [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.driver.spawn(context, instance, image_meta, [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] vm_ref = self.build_virtual_machine(instance, [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.634220] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] for vif in network_info: [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self._sync_wrapper(fn, *args, **kwargs) [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.wait() [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self[:] = self._gt.wait() [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self._exit_event.wait() [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] result = hub.switch() [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.634568] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return self.greenlet.switch() [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] result = function(*args, **kwargs) [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] return func(*args, **kwargs) [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise e [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] nwinfo = self.network_api.allocate_for_instance( [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] created_port_ids = self._update_ports_for_instance( [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] with excutils.save_and_reraise_exception(): [ 547.634980] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] self.force_reraise() [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise self.value [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] updated_port = self._update_port( [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] _ensure_no_port_binding_failure(port) [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] raise exception.PortBindingFailed(port_id=port['id']) [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] nova.exception.PortBindingFailed: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. [ 547.635330] env[62000]: ERROR nova.compute.manager [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] [ 547.635935] env[62000]: DEBUG nova.compute.utils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 547.636190] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.774s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.638118] env[62000]: INFO nova.compute.claims [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 547.644668] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Build of instance 34097eee-1eab-4b71-a450-8d8dddd94276 was re-scheduled: Binding failed for port 70f47b70-cc2d-4cb1-a9e4-63bd02b4407e, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 547.646314] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 547.646314] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquiring lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.646314] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Acquired lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.646314] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.682434] env[62000]: DEBUG nova.network.neutron [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.736040] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.041027] env[62000]: DEBUG nova.network.neutron [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.115182] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.185185] env[62000]: DEBUG oslo_concurrency.lockutils [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] Releasing lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.185457] env[62000]: DEBUG nova.compute.manager [req-9123f00b-7125-44f6-b06e-206f995555a7 req-ecafcbe4-7db9-4947-8f5c-56fad0aacb01 service nova] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Received event network-vif-deleted-9fc62892-3187-47ea-b8d5-13a99d49ddfb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.204574] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.437698] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.546335] env[62000]: INFO nova.compute.manager [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] [instance: e57fb81c-3337-4500-863c-d27eaf763caa] Took 1.08 seconds to deallocate network for instance. [ 548.618324] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.618759] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.620024] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 548.620024] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-127d2c9e-83aa-4314-8390-0cdbc30df966 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.631493] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc64efcd-cf62-4192-974f-2a3712fd14e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.660032] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc980828-985b-4da1-b3fe-e3cda8b84a6a could not be found. [ 548.660032] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 548.660032] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 548.660032] env[62000]: DEBUG oslo.service.loopingcall [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.661447] env[62000]: DEBUG nova.compute.manager [-] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.661447] env[62000]: DEBUG nova.network.neutron [-] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.709823] env[62000]: DEBUG nova.network.neutron [-] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.890779] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459e9bc0-0f70-4f9f-b536-474e168311f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.902141] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42e3fd71-f8cd-49c0-a991-181a276de2af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.938812] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056ac6a1-1c9e-4725-9375-f54b708a23b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.941645] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Releasing lock "refresh_cache-34097eee-1eab-4b71-a450-8d8dddd94276" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.942790] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 548.942790] env[62000]: DEBUG nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.942790] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.945343] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 548.945343] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.945343] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.945343] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.945343] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.945343] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.945343] env[62000]: ERROR nova.compute.manager raise self.value [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.945343] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 548.945343] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.945343] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 548.945745] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.945745] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 548.945745] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 548.945745] env[62000]: ERROR nova.compute.manager [ 548.945745] env[62000]: Traceback (most recent call last): [ 548.945745] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 548.945745] env[62000]: listener.cb(fileno) [ 548.945745] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.945745] env[62000]: result = function(*args, **kwargs) [ 548.945745] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.945745] env[62000]: return func(*args, **kwargs) [ 548.945745] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.945745] env[62000]: raise e [ 548.945745] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.945745] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 548.945745] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.945745] env[62000]: created_port_ids = self._update_ports_for_instance( [ 548.945745] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.945745] env[62000]: with excutils.save_and_reraise_exception(): [ 548.945745] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.945745] env[62000]: self.force_reraise() [ 548.945745] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.945745] env[62000]: raise self.value [ 548.945745] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.945745] env[62000]: updated_port = self._update_port( [ 548.945745] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.945745] env[62000]: _ensure_no_port_binding_failure(port) [ 548.945745] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.945745] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 548.946558] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 548.946558] env[62000]: Removing descriptor: 16 [ 548.946558] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] Traceback (most recent call last): [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] yield resources [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.driver.spawn(context, instance, image_meta, [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.946558] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] vm_ref = self.build_virtual_machine(instance, [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] for vif in network_info: [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self._sync_wrapper(fn, *args, **kwargs) [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.wait() [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self[:] = self._gt.wait() [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self._exit_event.wait() [ 548.946897] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] result = hub.switch() [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self.greenlet.switch() [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] result = function(*args, **kwargs) [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return func(*args, **kwargs) [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise e [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] nwinfo = self.network_api.allocate_for_instance( [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 548.947715] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] created_port_ids = self._update_ports_for_instance( [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] with excutils.save_and_reraise_exception(): [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.force_reraise() [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise self.value [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] updated_port = self._update_port( [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] _ensure_no_port_binding_failure(port) [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.948096] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise exception.PortBindingFailed(port_id=port['id']) [ 548.948506] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 548.948506] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] [ 548.948506] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Terminating instance [ 548.952180] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.952180] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.952180] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 548.954113] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6ef3c0a-a0a7-4266-a1c7-3138f14561d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.969987] env[62000]: DEBUG nova.compute.provider_tree [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.972223] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.009100] env[62000]: DEBUG nova.compute.manager [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Received event network-changed-39817f72-1f81-4f77-8c09-d9cf7fde1df4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 549.009100] env[62000]: DEBUG nova.compute.manager [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Refreshing instance network info cache due to event network-changed-39817f72-1f81-4f77-8c09-d9cf7fde1df4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 549.009100] env[62000]: DEBUG oslo_concurrency.lockutils [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] Acquiring lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.009581] env[62000]: DEBUG oslo_concurrency.lockutils [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] Acquired lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.011173] env[62000]: DEBUG nova.network.neutron [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Refreshing network info cache for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 549.214174] env[62000]: DEBUG nova.network.neutron [-] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.475923] env[62000]: DEBUG nova.network.neutron [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.482671] env[62000]: DEBUG nova.scheduler.client.report [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 549.511571] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.571192] env[62000]: DEBUG nova.network.neutron [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.590433] env[62000]: INFO nova.scheduler.client.report [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Deleted allocations for instance e57fb81c-3337-4500-863c-d27eaf763caa [ 549.719224] env[62000]: INFO nova.compute.manager [-] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Took 1.06 seconds to deallocate network for instance. [ 549.727489] env[62000]: DEBUG nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 549.727671] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.990138] env[62000]: INFO nova.compute.manager [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] [instance: 34097eee-1eab-4b71-a450-8d8dddd94276] Took 1.05 seconds to deallocate network for instance. [ 549.994484] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.995522] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 550.001089] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.948s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.003018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.003018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 550.003018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.015s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.010624] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11772571-e648-436f-982d-5f2cfc9c9b30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.022509] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c32c5b-57bf-4773-81ef-dd1f74c2fe4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.038685] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71f28ab0-3ac8-4f73-b77a-99325f1deef2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.051901] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2b7bce-b6d5-4b6f-bc25-9b9afddf44ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.092274] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181659MB free_disk=53GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 550.092353] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.093702] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.100080] env[62000]: DEBUG nova.network.neutron [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.102288] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f31c4b6e-d7e6-4fcc-a7f6-0304f21e5aa8 tempest-InstanceActionsTestJSON-1919170624 tempest-InstanceActionsTestJSON-1919170624-project-member] Lock "e57fb81c-3337-4500-863c-d27eaf763caa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.210s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.245425] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 550.245425] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.245425] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.245425] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.245425] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.245425] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.245425] env[62000]: ERROR nova.compute.manager raise self.value [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.245425] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 550.245425] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.245425] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 550.246101] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.246101] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 550.246101] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 550.246101] env[62000]: ERROR nova.compute.manager [ 550.246101] env[62000]: Traceback (most recent call last): [ 550.246101] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 550.246101] env[62000]: listener.cb(fileno) [ 550.246101] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.246101] env[62000]: result = function(*args, **kwargs) [ 550.246101] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.246101] env[62000]: return func(*args, **kwargs) [ 550.246101] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.246101] env[62000]: raise e [ 550.246101] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.246101] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 550.246101] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.246101] env[62000]: created_port_ids = self._update_ports_for_instance( [ 550.246101] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.246101] env[62000]: with excutils.save_and_reraise_exception(): [ 550.246101] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.246101] env[62000]: self.force_reraise() [ 550.246101] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.246101] env[62000]: raise self.value [ 550.246101] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.246101] env[62000]: updated_port = self._update_port( [ 550.246101] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.246101] env[62000]: _ensure_no_port_binding_failure(port) [ 550.246101] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.246101] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 550.246881] env[62000]: nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 550.246881] env[62000]: Removing descriptor: 14 [ 550.246881] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] Traceback (most recent call last): [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] yield resources [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.driver.spawn(context, instance, image_meta, [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.246881] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] vm_ref = self.build_virtual_machine(instance, [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] for vif in network_info: [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self._sync_wrapper(fn, *args, **kwargs) [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.wait() [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self[:] = self._gt.wait() [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self._exit_event.wait() [ 550.247278] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] result = hub.switch() [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self.greenlet.switch() [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] result = function(*args, **kwargs) [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return func(*args, **kwargs) [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise e [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] nwinfo = self.network_api.allocate_for_instance( [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 550.247610] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] created_port_ids = self._update_ports_for_instance( [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] with excutils.save_and_reraise_exception(): [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.force_reraise() [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise self.value [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] updated_port = self._update_port( [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] _ensure_no_port_binding_failure(port) [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.247935] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise exception.PortBindingFailed(port_id=port['id']) [ 550.248232] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 550.248232] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] [ 550.248232] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Terminating instance [ 550.253976] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.253976] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.253976] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 550.516479] env[62000]: DEBUG nova.compute.utils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 550.516591] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 550.598460] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.598853] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.599058] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 550.599742] env[62000]: DEBUG oslo_concurrency.lockutils [req-3438caad-c91c-4f9f-a8d4-dd43aaafef4d req-26bccd46-1d99-4446-990f-9b8c81b8c16a service nova] Releasing lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.600094] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d85f98f1-c395-44f2-96a0-2c0502673f19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.605047] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 550.615273] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab38cc94-7653-46e1-9585-af6e0072337a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.636901] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e99bfd97-1952-4166-aa19-560053bae307 could not be found. [ 550.637202] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 550.637396] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.637632] env[62000]: DEBUG oslo.service.loopingcall [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.637844] env[62000]: DEBUG nova.compute.manager [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.637943] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.662180] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.728607] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34484f14-b4f2-4473-ae6f-93031947cda7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.737302] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08a0e03a-fc79-401b-b2a7-e34c88f444d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.778439] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cffbf3b4-bcd3-43ed-9d66-84a058676fa6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.786746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-712e7daa-687e-47aa-b632-ff82d0f8aa8d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.801241] env[62000]: DEBUG nova.compute.provider_tree [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.810201] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.874243] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquiring lock "465d94c2-5ebf-4b05-966e-9272e2ddebb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.874517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "465d94c2-5ebf-4b05-966e-9272e2ddebb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.018282] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 551.038643] env[62000]: INFO nova.scheduler.client.report [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Deleted allocations for instance 34097eee-1eab-4b71-a450-8d8dddd94276 [ 551.054335] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.138054] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.168057] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.305820] env[62000]: DEBUG nova.scheduler.client.report [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 551.500422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "d7c25703-df04-4289-92d6-226180c43082" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.500800] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "d7c25703-df04-4289-92d6-226180c43082" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.549976] env[62000]: DEBUG oslo_concurrency.lockutils [None req-dd16599b-0df5-4139-82e3-50199a4a35d0 tempest-ServersAdminNegativeTestJSON-1057556697 tempest-ServersAdminNegativeTestJSON-1057556697-project-member] Lock "34097eee-1eab-4b71-a450-8d8dddd94276" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.754s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.558305] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 551.558767] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 551.558958] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 551.560118] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f25d2c0f-cad8-436b-a553-e11d40267c14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.573554] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8555add-e6f3-439e-a549-d715231ed328 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.604416] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa76f279-9efd-401f-823e-806425ce39e5 could not be found. [ 551.604416] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 551.604416] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 551.604644] env[62000]: DEBUG oslo.service.loopingcall [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 551.606161] env[62000]: DEBUG nova.compute.manager [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 551.606161] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 551.648147] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 551.667963] env[62000]: INFO nova.compute.manager [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Took 1.03 seconds to deallocate network for instance. [ 551.672607] env[62000]: DEBUG nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 551.673578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.810885] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.808s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.810885] env[62000]: ERROR nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Traceback (most recent call last): [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.driver.spawn(context, instance, image_meta, [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.810885] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] vm_ref = self.build_virtual_machine(instance, [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] for vif in network_info: [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self._sync_wrapper(fn, *args, **kwargs) [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.wait() [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self[:] = self._gt.wait() [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self._exit_event.wait() [ 551.811236] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] result = hub.switch() [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return self.greenlet.switch() [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] result = function(*args, **kwargs) [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] return func(*args, **kwargs) [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise e [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] nwinfo = self.network_api.allocate_for_instance( [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.811599] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] created_port_ids = self._update_ports_for_instance( [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] with excutils.save_and_reraise_exception(): [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] self.force_reraise() [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise self.value [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] updated_port = self._update_port( [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] _ensure_no_port_binding_failure(port) [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.812072] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] raise exception.PortBindingFailed(port_id=port['id']) [ 551.812410] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] nova.exception.PortBindingFailed: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. [ 551.812410] env[62000]: ERROR nova.compute.manager [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] [ 551.812410] env[62000]: DEBUG nova.compute.utils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 551.813664] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.308s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.814621] env[62000]: INFO nova.compute.claims [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 551.819339] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Build of instance b8b1038d-86b5-40ce-bc0c-4718013d8e73 was re-scheduled: Binding failed for port 220d283b-60ac-4137-aeaf-43e249ee516f, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 551.819661] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 551.819903] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.820065] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquired lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.820844] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 552.033284] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 552.053286] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 552.070273] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 552.070895] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 552.070895] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 552.070895] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 552.070895] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 552.071206] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 552.071688] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 552.071903] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 552.072038] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 552.072304] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 552.073121] env[62000]: DEBUG nova.virt.hardware [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 552.073831] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9f8432e-fafa-4004-8474-3c6ff15d5c20 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.084551] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae799235-36d9-4a93-8432-25cabc71c60e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.102483] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 552.111787] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 552.114244] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eae06d98-e992-442b-905a-c8d3fbec7e6b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.128026] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created folder: OpenStack in parent group-v4. [ 552.128272] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating folder: Project (777e25b48b2a42e388177e70a0145551). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 552.128961] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-914043dd-0432-43d8-93db-31c19ad642a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.140288] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created folder: Project (777e25b48b2a42e388177e70a0145551) in parent group-v201431. [ 552.140521] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating folder: Instances. Parent ref: group-v201432. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 552.140760] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f38a3645-e008-4e04-ae92-026ddede247b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.150229] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created folder: Instances in parent group-v201432. [ 552.150486] env[62000]: DEBUG oslo.service.loopingcall [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 552.150670] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 552.150878] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d7aa1c8-7846-47b5-b403-20c1315aed46 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.164284] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.171628] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 552.171628] env[62000]: value = "task-881808" [ 552.171628] env[62000]: _type = "Task" [ 552.171628] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.180459] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881808, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.371858] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.476450] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.578965] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.615600] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "a36de4b6-6928-4110-bc27-825ae58b15e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.615833] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "a36de4b6-6928-4110-bc27-825ae58b15e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 552.669056] env[62000]: INFO nova.compute.manager [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Took 1.06 seconds to deallocate network for instance. [ 552.671674] env[62000]: DEBUG nova.compute.claims [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.671871] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.686709] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881808, 'name': CreateVM_Task, 'duration_secs': 0.333321} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.686709] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 552.687694] env[62000]: DEBUG oslo_vmware.service [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132e75ad-764b-4756-9a11-44a953ec70f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.695983] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.696161] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.697467] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 552.698083] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9163c700-ad45-40e6-ab62-00ff7b56a0ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.704552] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 552.704552] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f293f-d966-cc5f-bd6a-7c151e466187" [ 552.704552] env[62000]: _type = "Task" [ 552.704552] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.716445] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f293f-d966-cc5f-bd6a-7c151e466187, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.979316] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Releasing lock "refresh_cache-b8b1038d-86b5-40ce-bc0c-4718013d8e73" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.979553] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.979721] env[62000]: DEBUG nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.979892] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 553.058729] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99ed900b-f050-4c50-937b-179fa828fbff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.066606] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec88f0db-7100-41e3-9576-b272fc4ef590 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.102554] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de65a668-1f02-4b49-bdbf-e59e17ed579c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.107950] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.112476] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9239b674-0b51-4c01-ac78-feeca63274ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.128711] env[62000]: DEBUG nova.compute.provider_tree [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.216498] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.216785] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 553.217235] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.220939] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.220939] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.220939] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da0bdf2b-15b5-4d64-8731-978da45b12e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.226110] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.226263] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 553.227018] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa6f0b62-4fb9-4520-b649-119b19ea59c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.234529] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bf9112a-ac36-4220-95cf-41a510e6cde0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.239209] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 553.239209] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c37f2-22f0-32c9-8090-5deb007f8d67" [ 553.239209] env[62000]: _type = "Task" [ 553.239209] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.247336] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c37f2-22f0-32c9-8090-5deb007f8d67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.614354] env[62000]: DEBUG nova.network.neutron [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.631714] env[62000]: DEBUG nova.scheduler.client.report [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.750398] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 553.750398] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating directory with path [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 553.750398] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3d0ea0e-0f5e-4652-9137-f34a356222aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.761489] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created directory with path [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 553.761711] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Fetch image to [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 553.762052] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Downloading image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk on the data store datastore1 {{(pid=62000) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 553.762657] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b97300-afb8-47c9-bd4a-865dff83d6a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.769776] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff01091-d78b-4c51-a216-6bbe128e584e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.779138] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d33ce7e-9877-4e4d-9667-74ebad3c3c77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.810725] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-953b9e6f-cfe3-4aab-ab3f-00a805422c7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.816786] env[62000]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-5e42da16-f55b-455e-99d1-13588c1f6e70 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.847225] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Downloading image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to the data store datastore1 {{(pid=62000) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 553.920605] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 554.118637] env[62000]: INFO nova.compute.manager [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: b8b1038d-86b5-40ce-bc0c-4718013d8e73] Took 1.14 seconds to deallocate network for instance. [ 554.136449] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.137016] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 554.141402] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.453s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.197244] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquiring lock "1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.197660] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.395731] env[62000]: DEBUG nova.compute.manager [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Received event network-changed-5fe122d3-6f5e-40c9-b7e5-31fd07e29683 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 554.395855] env[62000]: DEBUG nova.compute.manager [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Refreshing instance network info cache due to event network-changed-5fe122d3-6f5e-40c9-b7e5-31fd07e29683. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 554.397099] env[62000]: DEBUG oslo_concurrency.lockutils [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] Acquiring lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.397248] env[62000]: DEBUG oslo_concurrency.lockutils [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] Acquired lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.397414] env[62000]: DEBUG nova.network.neutron [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Refreshing network info cache for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 554.647653] env[62000]: DEBUG nova.compute.utils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 554.654397] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 554.654916] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 554.664979] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 554.667235] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 554.667235] env[62000]: DEBUG nova.compute.manager [req-4fb87c5e-af5e-4209-ba1f-a375ed0098f5 req-c49430e4-8670-4d76-892b-803b1c6423e2 service nova] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Received event network-vif-deleted-39817f72-1f81-4f77-8c09-d9cf7fde1df4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 554.719697] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Downloaded image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk on the data store datastore1 {{(pid=62000) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 554.723904] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 554.723904] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copying Virtual Disk [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk to [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 554.723904] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae186229-a093-49e2-bf5f-2ff456643bf5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.735536] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 554.735536] env[62000]: value = "task-881809" [ 554.735536] env[62000]: _type = "Task" [ 554.735536] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.741607] env[62000]: DEBUG nova.policy [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e22b1b4eb25c48678da44f35012a0bdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f770323d2c284f7a827579627c757f2f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.747118] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881809, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.942600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b77fa27c-311b-4383-9a66-4f886e4ddd49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.951707] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc79244-3c61-4cbf-96a4-1a082e7a5a98 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.989025] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6023f382-ed11-47c0-8270-f5dcd80cd067 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.996482] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ebc8bf-5f9b-4b03-9f54-f9bfed469cd7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.015440] env[62000]: DEBUG nova.compute.provider_tree [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.096722] env[62000]: DEBUG nova.network.neutron [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.151788] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 555.226311] env[62000]: INFO nova.scheduler.client.report [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Deleted allocations for instance b8b1038d-86b5-40ce-bc0c-4718013d8e73 [ 555.254389] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881809, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.407121] env[62000]: DEBUG nova.network.neutron [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 555.525214] env[62000]: DEBUG nova.scheduler.client.report [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.749367] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b3a83b40-f912-423e-be8a-3fae478e5a5c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "b8b1038d-86b5-40ce-bc0c-4718013d8e73" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.887s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.756205] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881809, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643319} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.756429] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copied Virtual Disk [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk to [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 555.756638] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleting the datastore file [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 555.757920] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ade3a742-6cad-49cd-b4af-3b52808bc100 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.764932] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 555.764932] env[62000]: value = "task-881810" [ 555.764932] env[62000]: _type = "Task" [ 555.764932] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.778252] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881810, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.912151] env[62000]: DEBUG oslo_concurrency.lockutils [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] Releasing lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.912255] env[62000]: DEBUG nova.compute.manager [req-0256bc66-1b84-4f07-98d0-daef382d70cf req-8c91ebef-3a12-4bda-a8e8-f81c63600f25 service nova] [instance: e99bfd97-1952-4166-aa19-560053bae307] Received event network-vif-deleted-5fe122d3-6f5e-40c9-b7e5-31fd07e29683 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 555.923967] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Successfully created port: 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 556.032040] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.890s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.032662] env[62000]: ERROR nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] Traceback (most recent call last): [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.driver.spawn(context, instance, image_meta, [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self._vmops.spawn(context, instance, image_meta, injected_files, [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] vm_ref = self.build_virtual_machine(instance, [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] vif_infos = vmwarevif.get_vif_info(self._session, [ 556.032662] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] for vif in network_info: [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self._sync_wrapper(fn, *args, **kwargs) [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.wait() [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self[:] = self._gt.wait() [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self._exit_event.wait() [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] result = hub.switch() [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 556.033042] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return self.greenlet.switch() [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] result = function(*args, **kwargs) [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] return func(*args, **kwargs) [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise e [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] nwinfo = self.network_api.allocate_for_instance( [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] created_port_ids = self._update_ports_for_instance( [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] with excutils.save_and_reraise_exception(): [ 556.033378] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] self.force_reraise() [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise self.value [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] updated_port = self._update_port( [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] _ensure_no_port_binding_failure(port) [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] raise exception.PortBindingFailed(port_id=port['id']) [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] nova.exception.PortBindingFailed: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. [ 556.033715] env[62000]: ERROR nova.compute.manager [instance: 4489993f-7d78-4891-8282-5d06d7689447] [ 556.033989] env[62000]: DEBUG nova.compute.utils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 556.035131] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.895s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.041207] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Build of instance 4489993f-7d78-4891-8282-5d06d7689447 was re-scheduled: Binding failed for port 15d1f9f0-7e00-4389-b925-9af9d7b68ef9, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 556.041207] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 556.041207] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquiring lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.041207] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Acquired lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.041453] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 556.165728] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 556.192321] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 556.192856] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 556.193039] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 556.193276] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 556.193463] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 556.193644] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 556.193934] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 556.194145] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 556.194351] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 556.194551] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 556.194763] env[62000]: DEBUG nova.virt.hardware [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 556.195670] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2cb110-a4c1-4ef9-9114-73961e7a476a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.205786] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89f1eb4d-90ce-44e3-b710-3becc2250e70 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.251726] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 556.275616] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881810, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025075} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.275879] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 556.276104] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Moving file from [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72. {{(pid=62000) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 556.276358] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-42ee5392-6ee5-412b-96b7-bbb3c7b67c68 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.285368] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 556.285368] env[62000]: value = "task-881811" [ 556.285368] env[62000]: _type = "Task" [ 556.285368] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.296893] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881811, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.491129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquiring lock "3bd34ec2-5253-4578-baf3-11775de18ed4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.492210] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "3bd34ec2-5253-4578-baf3-11775de18ed4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.579679] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.796828] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.803960] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881811, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024897} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.810094] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] File moved {{(pid=62000) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 556.810094] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Cleaning up location [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 556.810094] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleting the datastore file [datastore1] vmware_temp/fdc9fb42-e6d4-4893-ab9b-4bcce1745ec8 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 556.810094] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ccbef81e-7e5f-464c-8c34-a60208bb185e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.813048] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.822267] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 556.822267] env[62000]: value = "task-881812" [ 556.822267] env[62000]: _type = "Task" [ 556.822267] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.833753] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.881346] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2f6883-5806-45f9-9e2d-ff2a93c056af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.894494] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a08db14-a6b6-450d-9cc4-30aa85ad9592 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.936242] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3f36ed8-2af0-49bd-8a08-0c84b9054dde {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.944662] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375b848e-527d-449d-ac67-cb15719a668b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.959538] env[62000]: DEBUG nova.compute.provider_tree [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.026655] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquiring lock "6ac79769-fc49-4dd9-abae-fc2113983257" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.026954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "6ac79769-fc49-4dd9-abae-fc2113983257" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.230418] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "123175a1-51a6-4463-8d90-c6b10770f905" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.231088] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "123175a1-51a6-4463-8d90-c6b10770f905" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.310032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Releasing lock "refresh_cache-4489993f-7d78-4891-8282-5d06d7689447" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.310188] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 557.310389] env[62000]: DEBUG nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.310559] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 557.334550] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025789} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.334790] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.335605] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c06cbe6e-84be-4da9-b3f8-54bfe4ade8fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.340728] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 557.340728] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6b1c9-69f1-cd26-9ee8-c22181524fd2" [ 557.340728] env[62000]: _type = "Task" [ 557.340728] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.348877] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6b1c9-69f1-cd26-9ee8-c22181524fd2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.357973] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "e7739874-2088-48e3-b970-96b3d935477b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.358222] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "e7739874-2088-48e3-b970-96b3d935477b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.363443] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.462734] env[62000]: DEBUG nova.scheduler.client.report [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.828116] env[62000]: ERROR nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 557.828116] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.828116] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.828116] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.828116] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.828116] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.828116] env[62000]: ERROR nova.compute.manager raise self.value [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.828116] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.828116] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.828116] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.828782] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.828782] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.828782] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 557.828782] env[62000]: ERROR nova.compute.manager [ 557.828782] env[62000]: Traceback (most recent call last): [ 557.828782] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.828782] env[62000]: listener.cb(fileno) [ 557.828782] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.828782] env[62000]: result = function(*args, **kwargs) [ 557.828782] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.828782] env[62000]: return func(*args, **kwargs) [ 557.828782] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.828782] env[62000]: raise e [ 557.828782] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.828782] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 557.828782] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.828782] env[62000]: created_port_ids = self._update_ports_for_instance( [ 557.828782] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.828782] env[62000]: with excutils.save_and_reraise_exception(): [ 557.828782] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.828782] env[62000]: self.force_reraise() [ 557.828782] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.828782] env[62000]: raise self.value [ 557.828782] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.828782] env[62000]: updated_port = self._update_port( [ 557.828782] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.828782] env[62000]: _ensure_no_port_binding_failure(port) [ 557.828782] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.828782] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.829564] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 557.829564] env[62000]: Removing descriptor: 14 [ 557.829564] env[62000]: ERROR nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Traceback (most recent call last): [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] yield resources [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.driver.spawn(context, instance, image_meta, [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.829564] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] vm_ref = self.build_virtual_machine(instance, [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] for vif in network_info: [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self._sync_wrapper(fn, *args, **kwargs) [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.wait() [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self[:] = self._gt.wait() [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self._exit_event.wait() [ 557.829944] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] result = hub.switch() [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self.greenlet.switch() [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] result = function(*args, **kwargs) [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return func(*args, **kwargs) [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise e [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] nwinfo = self.network_api.allocate_for_instance( [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.830303] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] created_port_ids = self._update_ports_for_instance( [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] with excutils.save_and_reraise_exception(): [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.force_reraise() [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise self.value [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] updated_port = self._update_port( [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] _ensure_no_port_binding_failure(port) [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.830715] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise exception.PortBindingFailed(port_id=port['id']) [ 557.831029] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 557.831029] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] [ 557.831029] env[62000]: INFO nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Terminating instance [ 557.836025] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquiring lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.836025] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquired lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.836025] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.860308] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6b1c9-69f1-cd26-9ee8-c22181524fd2, 'name': SearchDatastore_Task, 'duration_secs': 0.012354} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.860496] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.860732] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 557.861293] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9754654b-6e87-4a7e-8bdf-76c95c50c9fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.866794] env[62000]: DEBUG nova.network.neutron [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.873676] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 557.873676] env[62000]: value = "task-881813" [ 557.873676] env[62000]: _type = "Task" [ 557.873676] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.888445] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881813, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.971014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.971014] env[62000]: ERROR nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Traceback (most recent call last): [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.driver.spawn(context, instance, image_meta, [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.971014] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] vm_ref = self.build_virtual_machine(instance, [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] for vif in network_info: [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self._sync_wrapper(fn, *args, **kwargs) [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.wait() [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self[:] = self._gt.wait() [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self._exit_event.wait() [ 557.971382] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] result = hub.switch() [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return self.greenlet.switch() [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] result = function(*args, **kwargs) [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] return func(*args, **kwargs) [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise e [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] nwinfo = self.network_api.allocate_for_instance( [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.971792] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] created_port_ids = self._update_ports_for_instance( [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] with excutils.save_and_reraise_exception(): [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] self.force_reraise() [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise self.value [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] updated_port = self._update_port( [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] _ensure_no_port_binding_failure(port) [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.972194] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] raise exception.PortBindingFailed(port_id=port['id']) [ 557.972486] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] nova.exception.PortBindingFailed: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. [ 557.972486] env[62000]: ERROR nova.compute.manager [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] [ 557.972486] env[62000]: DEBUG nova.compute.utils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 557.973613] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 8.246s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.978328] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Build of instance 6a67dde1-ca25-493e-8d65-8d4cea487c85 was re-scheduled: Binding failed for port 9fc62892-3187-47ea-b8d5-13a99d49ddfb, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 557.979993] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 557.979993] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquiring lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.979993] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Acquired lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.979993] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 558.285294] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquiring lock "b6186f74-8a8e-4145-8126-7048ad29673f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.285640] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "b6186f74-8a8e-4145-8126-7048ad29673f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.369947] env[62000]: INFO nova.compute.manager [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] [instance: 4489993f-7d78-4891-8282-5d06d7689447] Took 1.06 seconds to deallocate network for instance. [ 558.384819] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.389995] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881813, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.550444] env[62000]: DEBUG nova.compute.manager [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Received event network-changed-fa404718-9b22-4a2f-a565-58fd495c93fe {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 558.550444] env[62000]: DEBUG nova.compute.manager [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Refreshing instance network info cache due to event network-changed-fa404718-9b22-4a2f-a565-58fd495c93fe. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 558.550444] env[62000]: DEBUG oslo_concurrency.lockutils [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] Acquiring lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.550444] env[62000]: DEBUG oslo_concurrency.lockutils [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] Acquired lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.551384] env[62000]: DEBUG nova.network.neutron [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Refreshing network info cache for port fa404718-9b22-4a2f-a565-58fd495c93fe {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 558.710407] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.795470] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.815719] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a6f30c-602b-4e50-818f-3b68fceefc3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.822112] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.829429] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98b7dbd-2e76-470d-b882-83f0dd98e454 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.867491] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-320de1da-ce20-4f4d-bb87-8f7b1b145ae2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.876737] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34117883-971a-4f26-9201-6ce5c843ff76 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.898670] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 558.908977] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881813, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.52337} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.908977] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 558.908977] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 558.908977] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-051bfe0e-868f-43e9-8e89-4b314c4d7c3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.918747] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 558.918747] env[62000]: value = "task-881814" [ 558.918747] env[62000]: _type = "Task" [ 558.918747] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.928563] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881814, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.044359] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "7396796d-77e5-4364-b889-9cf3cbdad3c9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.044359] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "7396796d-77e5-4364-b889-9cf3cbdad3c9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 559.086842] env[62000]: DEBUG nova.network.neutron [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.299177] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Releasing lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.302058] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 559.302058] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 559.302058] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6a37f651-23be-458d-b093-77d457411b54 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.310698] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9db44e-067c-4e7f-a751-aa901aac238b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.323516] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Releasing lock "refresh_cache-6a67dde1-ca25-493e-8d65-8d4cea487c85" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.323951] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.323951] env[62000]: DEBUG nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.324175] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.334492] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance faa1794d-5941-40fb-bd78-2608fa2d5da0 could not be found. [ 559.334708] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 559.334890] env[62000]: INFO nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 559.335680] env[62000]: DEBUG oslo.service.loopingcall [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.335854] env[62000]: DEBUG nova.compute.manager [-] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.335949] env[62000]: DEBUG nova.network.neutron [-] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.377665] env[62000]: DEBUG nova.network.neutron [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.389671] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.401731] env[62000]: DEBUG nova.network.neutron [-] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.404192] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 559.432144] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881814, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068075} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.432465] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 559.434145] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c15c8f-3e78-4872-b906-922f62f76753 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.437781] env[62000]: INFO nova.scheduler.client.report [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Deleted allocations for instance 4489993f-7d78-4891-8282-5d06d7689447 [ 559.469028] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 559.469447] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b3164b90-cb85-4976-b3a1-a3fa701184c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.494050] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 559.494050] env[62000]: value = "task-881815" [ 559.494050] env[62000]: _type = "Task" [ 559.494050] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.503298] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881815, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.113333] env[62000]: DEBUG oslo_concurrency.lockutils [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] Releasing lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.113333] env[62000]: DEBUG nova.compute.manager [req-b741f305-1601-4dd9-ad89-0397592e9978 req-3c3afb50-c7ad-47e6-9840-ae597a838a0d service nova] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Received event network-vif-deleted-fa404718-9b22-4a2f-a565-58fd495c93fe {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 560.113333] env[62000]: DEBUG nova.network.neutron [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.113828] env[62000]: DEBUG nova.network.neutron [-] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.119188] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.144s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.119188] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Traceback (most recent call last): [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.driver.spawn(context, instance, image_meta, [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.119188] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] vm_ref = self.build_virtual_machine(instance, [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] for vif in network_info: [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self._sync_wrapper(fn, *args, **kwargs) [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.wait() [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self[:] = self._gt.wait() [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self._exit_event.wait() [ 560.119524] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] result = hub.switch() [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return self.greenlet.switch() [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] result = function(*args, **kwargs) [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] return func(*args, **kwargs) [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise e [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] nwinfo = self.network_api.allocate_for_instance( [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 560.119876] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] created_port_ids = self._update_ports_for_instance( [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] with excutils.save_and_reraise_exception(): [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] self.force_reraise() [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise self.value [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] updated_port = self._update_port( [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] _ensure_no_port_binding_failure(port) [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.120255] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] raise exception.PortBindingFailed(port_id=port['id']) [ 560.120579] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] nova.exception.PortBindingFailed: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. [ 560.120579] env[62000]: ERROR nova.compute.manager [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] [ 560.120579] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 560.120804] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c839ea78-01f5-4280-9a99-2db43da07173 tempest-FloatingIPsAssociationTestJSON-359561633 tempest-FloatingIPsAssociationTestJSON-359561633-project-member] Lock "4489993f-7d78-4891-8282-5d06d7689447" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.061s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.122105] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Build of instance dc980828-985b-4da1-b3fe-e3cda8b84a6a was re-scheduled: Binding failed for port 39817f72-1f81-4f77-8c09-d9cf7fde1df4, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 560.125032] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 560.125032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.125032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.125032] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 560.125032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.032s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.141661] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881815, 'name': ReconfigVM_Task, 'duration_secs': 0.264461} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.142709] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 560.143697] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6c13a7d-c869-485e-a7e7-47f54c9290c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.157842] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 560.157842] env[62000]: value = "task-881816" [ 560.157842] env[62000]: _type = "Task" [ 560.157842] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.183028] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881816, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.623633] env[62000]: INFO nova.compute.manager [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] [instance: 6a67dde1-ca25-493e-8d65-8d4cea487c85] Took 1.30 seconds to deallocate network for instance. [ 560.630242] env[62000]: INFO nova.compute.manager [-] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Took 1.29 seconds to deallocate network for instance. [ 560.631095] env[62000]: DEBUG nova.compute.claims [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 560.631289] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.650567] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.666997] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881816, 'name': Rename_Task, 'duration_secs': 0.140897} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.668669] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 560.669416] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c951c1f0-d110-41aa-a6d8-06d176d1ed94 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.674774] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.678539] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 560.678539] env[62000]: value = "task-881817" [ 560.678539] env[62000]: _type = "Task" [ 560.678539] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.688875] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.907393] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.192372] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.204034] env[62000]: DEBUG oslo_vmware.api [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881817, 'name': PowerOnVM_Task, 'duration_secs': 0.422239} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.204034] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 561.204034] env[62000]: INFO nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Took 9.17 seconds to spawn the instance on the hypervisor. [ 561.204034] env[62000]: DEBUG nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 561.204034] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee346d6-d27f-4da7-a8b5-7f3ac42bdfe9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.414851] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-dc980828-985b-4da1-b3fe-e3cda8b84a6a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.415164] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 561.417503] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 561.417691] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 561.437165] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 561.657458] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquiring lock "289e3403-c90d-40f0-9e18-1d10c0f8aec4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.657458] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "289e3403-c90d-40f0-9e18-1d10c0f8aec4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.673220] env[62000]: INFO nova.scheduler.client.report [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Deleted allocations for instance 6a67dde1-ca25-493e-8d65-8d4cea487c85 [ 561.688414] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 6a67dde1-ca25-493e-8d65-8d4cea487c85 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.733157] env[62000]: INFO nova.compute.manager [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Took 23.90 seconds to build instance. [ 561.945641] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 562.095857] env[62000]: DEBUG nova.compute.manager [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Received event network-changed-3c71f5dc-6fa7-46bb-84cb-a3c384fb0585 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 562.095857] env[62000]: DEBUG nova.compute.manager [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Refreshing instance network info cache due to event network-changed-3c71f5dc-6fa7-46bb-84cb-a3c384fb0585. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 562.095857] env[62000]: DEBUG oslo_concurrency.lockutils [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] Acquiring lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 562.095857] env[62000]: DEBUG oslo_concurrency.lockutils [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] Acquired lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.095857] env[62000]: DEBUG nova.network.neutron [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Refreshing network info cache for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 562.190678] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance dc980828-985b-4da1-b3fe-e3cda8b84a6a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.190846] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance e99bfd97-1952-4166-aa19-560053bae307 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 562.190971] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance fa76f279-9efd-401f-823e-806425ce39e5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 562.191098] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 15409fa7-254c-435e-8080-46d3f65b2d46 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 562.191209] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance faa1794d-5941-40fb-bd78-2608fa2d5da0 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 562.192701] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4ee059bf-11b8-400a-9f01-e662ff1a3687 tempest-ServerExternalEventsTest-639902409 tempest-ServerExternalEventsTest-639902409-project-member] Lock "6a67dde1-ca25-493e-8d65-8d4cea487c85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.641s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.238447] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ce4365c0-9881-4ec2-a516-bf2010b44e0b tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.413s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 562.447754] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: dc980828-985b-4da1-b3fe-e3cda8b84a6a] Took 1.03 seconds to deallocate network for instance. [ 562.632619] env[62000]: DEBUG nova.network.neutron [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 562.653307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "ab167fa4-062b-4137-801c-9b57a17b20e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 562.654030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "ab167fa4-062b-4137-801c-9b57a17b20e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 562.702617] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.707544] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 57cc0a22-28ff-4a24-9fea-5c6a0a699a67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.742661] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 562.757386] env[62000]: DEBUG nova.network.neutron [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.212975] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance c0f34d3e-39a1-4487-bd1f-ade41ef6832c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.239514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.260778] env[62000]: DEBUG oslo_concurrency.lockutils [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] Releasing lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.260778] env[62000]: DEBUG nova.compute.manager [req-571d3532-efb2-4175-8ac9-f2e57516d48d req-57549f84-f2f6-498e-97f6-0e80746c47ee service nova] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Received event network-vif-deleted-3c71f5dc-6fa7-46bb-84cb-a3c384fb0585 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 563.261928] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.499227] env[62000]: INFO nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Deleted allocations for instance dc980828-985b-4da1-b3fe-e3cda8b84a6a [ 563.718372] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 465d94c2-5ebf-4b05-966e-9272e2ddebb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.016415] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "dc980828-985b-4da1-b3fe-e3cda8b84a6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.372s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 564.205767] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "8a414b1b-4513-405a-a0f0-a0af61722e3d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.208398] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "8a414b1b-4513-405a-a0f0-a0af61722e3d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.223840] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance d7c25703-df04-4289-92d6-226180c43082 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.516423] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 564.585180] env[62000]: INFO nova.compute.manager [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Rebuilding instance [ 564.661138] env[62000]: DEBUG nova.compute.manager [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 564.662083] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177230c0-3c13-4397-b2e3-4117b6df2db3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.671485] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "5ef8c2a0-771f-4931-a183-440fd5d998fe" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.671581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "5ef8c2a0-771f-4931-a183-440fd5d998fe" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 564.727564] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance a36de4b6-6928-4110-bc27-825ae58b15e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.056277] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.178944] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 565.178944] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f2810932-25e5-4fac-b24f-a87103f7ff3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.190769] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 565.190769] env[62000]: value = "task-881818" [ 565.190769] env[62000]: _type = "Task" [ 565.190769] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.205962] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881818, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.236442] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.708012] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881818, 'name': PowerOffVM_Task, 'duration_secs': 0.185213} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.708012] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 565.708195] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 565.709008] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f81ac69a-576b-40e4-9431-b501cdcb69d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.717055] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 565.717313] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e9d0362-1432-4cd1-9ae4-1892419bf37e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.743784] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 3bd34ec2-5253-4578-baf3-11775de18ed4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 565.750483] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 565.750697] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 565.750874] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleting the datastore file [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 565.752752] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-662df59b-787e-44b1-9102-6cad06cafbca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.761998] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.762679] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.768902] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 565.768902] env[62000]: value = "task-881820" [ 565.768902] env[62000]: _type = "Task" [ 565.768902] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.782120] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881820, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.971329] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "042678e8-2028-416a-bc4a-c473c7f073bd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 565.971329] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "042678e8-2028-416a-bc4a-c473c7f073bd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.252587] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 6ac79769-fc49-4dd9-abae-fc2113983257 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 566.286893] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881820, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.17313} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.289846] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 566.289846] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 566.289846] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 566.387658] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 566.388350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.755694] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 123175a1-51a6-4463-8d90-c6b10770f905 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.263814] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance e7739874-2088-48e3-b970-96b3d935477b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.334380] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 567.334657] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 567.334955] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 567.335027] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 567.335153] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 567.335295] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 567.335501] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 567.335649] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 567.335810] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 567.335965] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 567.336149] env[62000]: DEBUG nova.virt.hardware [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 567.337094] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a1f30e6-8e6a-4e70-8a15-b5ce456a74f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.350226] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5c01f3f-eaca-4bc2-a917-b0ef76a00d35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.370472] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 567.379418] env[62000]: DEBUG oslo.service.loopingcall [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.379712] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 567.379922] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-77bc943d-7a27-433e-b38c-3253f7cad838 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.402309] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 567.402309] env[62000]: value = "task-881821" [ 567.402309] env[62000]: _type = "Task" [ 567.402309] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.411388] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881821, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.440433] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "f36c4c0a-6777-468f-8798-44093023b2d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 567.440658] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "f36c4c0a-6777-468f-8798-44093023b2d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.766880] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance b6186f74-8a8e-4145-8126-7048ad29673f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 567.915511] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881821, 'name': CreateVM_Task, 'duration_secs': 0.297274} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.915704] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 567.916286] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.916444] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.916758] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 567.917011] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ad2c640-369f-4617-a7ca-e2160a4c16d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.922641] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 567.922641] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529e0662-1b64-5312-134a-0183fb2865b5" [ 567.922641] env[62000]: _type = "Task" [ 567.922641] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.933717] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529e0662-1b64-5312-134a-0183fb2865b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.270482] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 7396796d-77e5-4364-b889-9cf3cbdad3c9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.434372] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529e0662-1b64-5312-134a-0183fb2865b5, 'name': SearchDatastore_Task, 'duration_secs': 0.011924} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.434924] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.435222] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 568.435509] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.435791] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.435873] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 568.436445] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d18e240d-abdd-4088-b7ae-c8042965a462 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.444733] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 568.445566] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 568.445724] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7d0e6bd-0ff1-4273-8e9d-7d29b7b0514c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.450783] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 568.450783] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523f5e4b-4610-7149-da6d-8f3e4d40fb8f" [ 568.450783] env[62000]: _type = "Task" [ 568.450783] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.461429] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523f5e4b-4610-7149-da6d-8f3e4d40fb8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.738552] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.738552] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.774832] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 289e3403-c90d-40f0-9e18-1d10c0f8aec4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 568.774832] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 568.774832] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 568.888372] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "e99bfd97-1952-4166-aa19-560053bae307" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.967573] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523f5e4b-4610-7149-da6d-8f3e4d40fb8f, 'name': SearchDatastore_Task, 'duration_secs': 0.009069} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.973076] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a70c028-5f9b-4575-ab37-2bf897f60262 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.987979] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 568.987979] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525ea835-f027-20a3-1aec-2ec6de645e10" [ 568.987979] env[62000]: _type = "Task" [ 568.987979] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.999747] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525ea835-f027-20a3-1aec-2ec6de645e10, 'name': SearchDatastore_Task, 'duration_secs': 0.008589} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.000638] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.000918] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 569.002500] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a693c3ca-2da1-4b54-a33d-16706bfc093c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.021021] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 569.021021] env[62000]: value = "task-881822" [ 569.021021] env[62000]: _type = "Task" [ 569.021021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.029579] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881822, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.307813] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d4d358-0ce8-4833-9071-83d10b9e8ff0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.319266] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316743f8-c402-4766-9d6e-a284ca02cd44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.362846] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c8223f-ec24-4b39-b3a6-91ef568451e1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.369307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "fa76f279-9efd-401f-823e-806425ce39e5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.373655] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84d53aee-1c42-4222-914e-1c731b06840c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.392263] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.529595] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881822, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493551} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.529595] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 569.529595] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 569.529595] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9c4c7ae-444a-4c94-8c11-3055974bec49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.537616] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 569.537616] env[62000]: value = "task-881823" [ 569.537616] env[62000]: _type = "Task" [ 569.537616] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.548805] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881823, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.895981] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.052342] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881823, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.103321} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.052742] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 570.053611] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-854e1b60-d1ad-4099-946e-a002f760a5b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.079061] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 570.079061] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5fc601f-c1f6-41da-9232-0ebb974eeabc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.102800] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 570.102800] env[62000]: value = "task-881824" [ 570.102800] env[62000]: _type = "Task" [ 570.102800] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.113174] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881824, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.403359] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 570.403359] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.278s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.403359] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.264s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.404026] env[62000]: INFO nova.compute.claims [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.615833] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881824, 'name': ReconfigVM_Task, 'duration_secs': 0.289196} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.616155] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 570.616898] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d01fff4f-c052-40ce-90ec-ad5d2e01cf71 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.624522] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 570.624522] env[62000]: value = "task-881825" [ 570.624522] env[62000]: _type = "Task" [ 570.624522] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.635178] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881825, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.139606] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881825, 'name': Rename_Task, 'duration_secs': 0.131117} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.139884] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 571.142575] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-31b0ce0e-aaf5-46f2-a4d3-ad660110fe09 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.149908] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 571.149908] env[62000]: value = "task-881826" [ 571.149908] env[62000]: _type = "Task" [ 571.149908] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.160623] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.353376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.353608] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.669359] env[62000]: DEBUG oslo_vmware.api [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881826, 'name': PowerOnVM_Task, 'duration_secs': 0.457754} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.669931] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 571.670450] env[62000]: DEBUG nova.compute.manager [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 571.671588] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f79c4f8-4df3-4eb5-9831-d7c7f1993d79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.959356] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f85123-9073-4636-ba42-93866f21ca13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.967350] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f22abc1-f224-4a12-83fe-8d6939fa8eeb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.011549] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd69aff-31b7-4755-8149-f492778feb0f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.022583] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e7987b-7762-4c47-9071-61e8d2e91f92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.039545] env[62000]: DEBUG nova.compute.provider_tree [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 572.200033] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.542567] env[62000]: DEBUG nova.scheduler.client.report [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.857594] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "762e429f-f690-43c4-95eb-877caf1cdad7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.857594] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.049485] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 573.050468] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 573.055085] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.382s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.562273] env[62000]: DEBUG nova.compute.utils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 573.566206] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 573.566952] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 573.857278] env[62000]: DEBUG nova.policy [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd099918d0cc948b987ef0b97b5c69ade', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cee595612b62474cb70ac65e436921c3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 574.076207] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 574.082579] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd51cee4-c95c-48fd-892b-488fa670c4ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.099632] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe3717d0-9135-40c4-8146-049200cb2bc4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.137481] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35631ab0-89a4-4eb6-98a7-a4b15c5194cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.146036] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f8dcff1-12f0-42b4-9dd1-4600cb58c96d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.169452] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.404534] env[62000]: INFO nova.compute.manager [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Rebuilding instance [ 574.456028] env[62000]: DEBUG nova.compute.manager [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 574.456601] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b709ec0-048f-418f-b686-77551524dfc3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.672905] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.972464] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 574.972464] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36627ed0-83f4-41ec-a35d-ec8727d1aaee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.979428] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 574.979428] env[62000]: value = "task-881827" [ 574.979428] env[62000]: _type = "Task" [ 574.979428] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 574.991018] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.101472] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 575.141857] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 575.141857] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 575.141857] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 575.142405] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 575.142649] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 575.142847] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 575.143568] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 575.143568] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 575.143568] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 575.143836] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 575.143836] env[62000]: DEBUG nova.virt.hardware [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 575.145837] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe4763d-f238-4bd0-be11-fb477a68607e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.156063] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c2eb6d-6f97-42b2-b184-b2ffe460f85e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.183014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.128s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.184102] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] Traceback (most recent call last): [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.driver.spawn(context, instance, image_meta, [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self._vmops.spawn(context, instance, image_meta, injected_files, [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] vm_ref = self.build_virtual_machine(instance, [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] vif_infos = vmwarevif.get_vif_info(self._session, [ 575.184102] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] for vif in network_info: [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self._sync_wrapper(fn, *args, **kwargs) [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.wait() [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self[:] = self._gt.wait() [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self._exit_event.wait() [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] result = hub.switch() [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 575.184509] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return self.greenlet.switch() [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] result = function(*args, **kwargs) [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] return func(*args, **kwargs) [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise e [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] nwinfo = self.network_api.allocate_for_instance( [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] created_port_ids = self._update_ports_for_instance( [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] with excutils.save_and_reraise_exception(): [ 575.184893] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] self.force_reraise() [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise self.value [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] updated_port = self._update_port( [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] _ensure_no_port_binding_failure(port) [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] raise exception.PortBindingFailed(port_id=port['id']) [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] nova.exception.PortBindingFailed: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. [ 575.185294] env[62000]: ERROR nova.compute.manager [instance: e99bfd97-1952-4166-aa19-560053bae307] [ 575.185587] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 575.185810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.607s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.187718] env[62000]: INFO nova.compute.claims [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.190032] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Build of instance e99bfd97-1952-4166-aa19-560053bae307 was re-scheduled: Binding failed for port 5fe122d3-6f5e-40c9-b7e5-31fd07e29683, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 575.190772] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 575.190772] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 575.190905] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 575.191053] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 575.288318] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Successfully created port: 4b926553-f0d0-4888-a629-b4a735daa258 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 575.492032] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881827, 'name': PowerOffVM_Task, 'duration_secs': 0.145027} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 575.492331] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 575.492547] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 575.493382] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fd91930-7735-4b29-b02a-80a5c702d59d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.500572] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 575.500824] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c06a38dd-9b48-4eb5-b583-bebc96c820d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.525424] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 575.526132] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 575.526345] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Deleting the datastore file [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 575.526685] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2743aafc-6483-4151-a164-a36183bd143f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.536866] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 575.536866] env[62000]: value = "task-881829" [ 575.536866] env[62000]: _type = "Task" [ 575.536866] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.546938] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881829, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.753160] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 576.051729] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881829, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.051871] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 576.052041] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 576.052240] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 576.206975] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.668451] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02fcfb3-842f-4acf-b3ee-3bcfdf8eef35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.676621] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9a9d4b-955b-42dd-880d-6b239d1bb400 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.713044] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.713284] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 576.713461] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 576.714857] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 576.716272] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7ee406-dab5-4f35-a0a2-b16f2574f9fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.724070] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-990e8c5a-90ac-4a6b-8006-e98ef1c58f86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.737743] env[62000]: DEBUG nova.compute.provider_tree [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.775614] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 577.100094] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.100094] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.100094] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.100094] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.100339] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.100339] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.100339] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.100339] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.100339] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.100788] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.101152] env[62000]: DEBUG nova.virt.hardware [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.102520] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9a97cf-dd7d-4a9a-985b-5c6dd5137d0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.112983] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bc9cd1-5d7a-44aa-9b0e-9180ffd83c50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.130971] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 577.135265] env[62000]: DEBUG oslo.service.loopingcall [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.135501] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 577.135750] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f9c9bd57-70c0-4984-b120-198851a49195 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.154103] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 577.154103] env[62000]: value = "task-881830" [ 577.154103] env[62000]: _type = "Task" [ 577.154103] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.162670] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881830, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.241112] env[62000]: DEBUG nova.scheduler.client.report [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.278653] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.670909] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881830, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.746736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.749042] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.754688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.083s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.787668] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Took 1.07 seconds to deallocate network for instance. [ 578.168298] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881830, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.258387] env[62000]: DEBUG nova.compute.utils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 578.261991] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 578.261991] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 578.452086] env[62000]: DEBUG nova.policy [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a55903eebb6c4da18b0eaff1f244d6f3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03ad93ef1d244ef2b8980357ca2abaa8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 578.669454] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881830, 'name': CreateVM_Task, 'duration_secs': 1.310188} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.669454] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 578.670632] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.671048] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.671609] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 578.671609] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66958853-0647-4065-8adc-c120fe68d458 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.678532] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 578.678532] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528cbf72-9e29-5984-21d8-a00072da174c" [ 578.678532] env[62000]: _type = "Task" [ 578.678532] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.686621] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528cbf72-9e29-5984-21d8-a00072da174c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.724608] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b703dbcf-c9f0-429f-8a07-9fd6d91415dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.739559] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4cea1c5-7aff-49d6-89b2-0b175597404f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.769735] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.773245] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1007ed5-fb9d-41c4-8b71-8a4fcb606840 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.782324] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de84a40-e17b-4c5d-8e0d-fb5b70afc8dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.800138] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 578.825934] env[62000]: INFO nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Deleted allocations for instance e99bfd97-1952-4166-aa19-560053bae307 [ 579.196637] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528cbf72-9e29-5984-21d8-a00072da174c, 'name': SearchDatastore_Task, 'duration_secs': 0.010657} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.196986] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 579.197287] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 579.197554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.197697] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.197872] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 579.198157] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac78677d-7d6b-4863-bde1-2eff50d4cf08 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.208835] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 579.209105] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 579.209925] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b46694b5-5e21-46d8-9c88-8eb0cd3ef6a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.217592] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 579.217592] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a3bfac-4bc9-febd-52cf-0f640263afc6" [ 579.217592] env[62000]: _type = "Task" [ 579.217592] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.228156] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a3bfac-4bc9-febd-52cf-0f640263afc6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.303175] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.338357] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.649s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.339564] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 10.451s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.339809] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "e99bfd97-1952-4166-aa19-560053bae307-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.340043] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.340208] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.343732] env[62000]: INFO nova.compute.manager [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Terminating instance [ 579.347965] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.348172] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.351101] env[62000]: DEBUG nova.network.neutron [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.455234] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Successfully created port: 16cacb86-1eef-4bc8-8bb8-19573360c52b {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 579.730044] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a3bfac-4bc9-febd-52cf-0f640263afc6, 'name': SearchDatastore_Task, 'duration_secs': 0.009498} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.731113] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f0602050-a83f-4f6b-99b0-2a7631d8f9fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.738024] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 579.738024] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52084867-ac47-230d-f4ae-82bcec1f938c" [ 579.738024] env[62000]: _type = "Task" [ 579.738024] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.747181] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52084867-ac47-230d-f4ae-82bcec1f938c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.783656] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 579.814714] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 579.815369] env[62000]: ERROR nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] Traceback (most recent call last): [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.driver.spawn(context, instance, image_meta, [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] vm_ref = self.build_virtual_machine(instance, [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.815369] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] for vif in network_info: [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self._sync_wrapper(fn, *args, **kwargs) [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.wait() [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self[:] = self._gt.wait() [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self._exit_event.wait() [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] result = hub.switch() [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.815784] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return self.greenlet.switch() [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] result = function(*args, **kwargs) [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] return func(*args, **kwargs) [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise e [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] nwinfo = self.network_api.allocate_for_instance( [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] created_port_ids = self._update_ports_for_instance( [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] with excutils.save_and_reraise_exception(): [ 579.816376] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] self.force_reraise() [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise self.value [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] updated_port = self._update_port( [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] _ensure_no_port_binding_failure(port) [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] raise exception.PortBindingFailed(port_id=port['id']) [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] nova.exception.PortBindingFailed: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. [ 579.817484] env[62000]: ERROR nova.compute.manager [instance: fa76f279-9efd-401f-823e-806425ce39e5] [ 579.817871] env[62000]: DEBUG nova.compute.utils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 579.819236] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 579.819453] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 579.824055] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 579.824055] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 579.824055] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 579.824055] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 579.824055] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 579.824486] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 579.824486] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 579.824486] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 579.824486] env[62000]: DEBUG nova.virt.hardware [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 579.824486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.009s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.824654] env[62000]: INFO nova.compute.claims [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 579.828046] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b68b05-0364-4498-81b4-d023d4a5d24d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.833643] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Build of instance fa76f279-9efd-401f-823e-806425ce39e5 was re-scheduled: Binding failed for port fa404718-9b22-4a2f-a565-58fd495c93fe, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 579.833716] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 579.834690] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.834690] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.834690] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 579.843052] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eca86ed-1ffe-4483-a0ae-7f5058e7b667 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.854510] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 579.910124] env[62000]: DEBUG nova.network.neutron [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.099220] env[62000]: ERROR nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 580.099220] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.099220] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.099220] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.099220] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.099220] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.099220] env[62000]: ERROR nova.compute.manager raise self.value [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.099220] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 580.099220] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.099220] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 580.099694] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.099694] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 580.099694] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 580.099694] env[62000]: ERROR nova.compute.manager [ 580.099694] env[62000]: Traceback (most recent call last): [ 580.099694] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 580.099694] env[62000]: listener.cb(fileno) [ 580.099694] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.099694] env[62000]: result = function(*args, **kwargs) [ 580.099694] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.099694] env[62000]: return func(*args, **kwargs) [ 580.099694] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.099694] env[62000]: raise e [ 580.099694] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.099694] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 580.099694] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.099694] env[62000]: created_port_ids = self._update_ports_for_instance( [ 580.099694] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.099694] env[62000]: with excutils.save_and_reraise_exception(): [ 580.099694] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.099694] env[62000]: self.force_reraise() [ 580.099694] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.099694] env[62000]: raise self.value [ 580.099694] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.099694] env[62000]: updated_port = self._update_port( [ 580.099694] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.099694] env[62000]: _ensure_no_port_binding_failure(port) [ 580.099694] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.099694] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 580.100576] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 580.100576] env[62000]: Removing descriptor: 19 [ 580.100576] env[62000]: ERROR nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Traceback (most recent call last): [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] yield resources [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.driver.spawn(context, instance, image_meta, [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 580.100576] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] vm_ref = self.build_virtual_machine(instance, [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] vif_infos = vmwarevif.get_vif_info(self._session, [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] for vif in network_info: [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self._sync_wrapper(fn, *args, **kwargs) [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.wait() [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self[:] = self._gt.wait() [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self._exit_event.wait() [ 580.100967] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] result = hub.switch() [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self.greenlet.switch() [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] result = function(*args, **kwargs) [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return func(*args, **kwargs) [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise e [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] nwinfo = self.network_api.allocate_for_instance( [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 580.101463] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] created_port_ids = self._update_ports_for_instance( [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] with excutils.save_and_reraise_exception(): [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.force_reraise() [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise self.value [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] updated_port = self._update_port( [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] _ensure_no_port_binding_failure(port) [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 580.102205] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise exception.PortBindingFailed(port_id=port['id']) [ 580.102749] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 580.102749] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] [ 580.102749] env[62000]: INFO nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Terminating instance [ 580.102749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquiring lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.102749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquired lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.102749] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 580.193623] env[62000]: DEBUG nova.network.neutron [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.251302] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52084867-ac47-230d-f4ae-82bcec1f938c, 'name': SearchDatastore_Task, 'duration_secs': 0.008921} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.251810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.252065] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 580.252331] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8b7a4972-f836-48e1-8b68-751071bc9c86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.259330] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 580.259330] env[62000]: value = "task-881831" [ 580.259330] env[62000]: _type = "Task" [ 580.259330] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.270858] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881831, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.341275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.341520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.375974] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.384041] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.388234] env[62000]: DEBUG nova.compute.manager [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Received event network-changed-4b926553-f0d0-4888-a629-b4a735daa258 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 580.388440] env[62000]: DEBUG nova.compute.manager [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Refreshing instance network info cache due to event network-changed-4b926553-f0d0-4888-a629-b4a735daa258. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 580.388870] env[62000]: DEBUG oslo_concurrency.lockutils [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] Acquiring lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 580.656046] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 580.699784] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-e99bfd97-1952-4166-aa19-560053bae307" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.703680] env[62000]: DEBUG nova.compute.manager [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.704034] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 580.704369] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51225178-7ae9-482f-8191-f55e09be3cad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.713346] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.716281] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.716281] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.725924] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b64718-f885-4f20-80fa-b8992598f00f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.755570] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e99bfd97-1952-4166-aa19-560053bae307 could not be found. [ 580.755795] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 580.755984] env[62000]: INFO nova.compute.manager [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: e99bfd97-1952-4166-aa19-560053bae307] Took 0.05 seconds to destroy the instance on the hypervisor. [ 580.756276] env[62000]: DEBUG oslo.service.loopingcall [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.756487] env[62000]: DEBUG nova.compute.manager [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.756610] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 580.769831] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881831, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49509} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.770448] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 580.770727] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 580.771307] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d521a555-db16-4412-9be0-2cdb19f79509 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.778062] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 580.778062] env[62000]: value = "task-881832" [ 580.778062] env[62000]: _type = "Task" [ 580.778062] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.786658] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881832, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.795297] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.012074] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.219775] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.220055] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 581.220289] env[62000]: DEBUG nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.220665] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.258692] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.287965] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881832, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071325} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.288257] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 581.293430] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ee0755-33ab-45f4-90cd-0915b2e92372 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.306573] env[62000]: DEBUG nova.network.neutron [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.315751] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfiguring VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 581.317600] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c54880aa-f292-4a3f-a027-ea4f45eed418 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.332254] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f323e226-69e0-4699-a5ee-f9f7e94ab672 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.340496] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae64b57-c40b-4479-a5a2-d76f03da8de1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.346804] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 581.346804] env[62000]: value = "task-881833" [ 581.346804] env[62000]: _type = "Task" [ 581.346804] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.374081] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7948b0cb-b718-461a-8fc9-8a06dcf366e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.379774] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.384507] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0c7ba5f-1a5d-4b27-a610-3b08c7c3efc1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.398788] env[62000]: DEBUG nova.compute.provider_tree [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.516602] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Releasing lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.517490] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 581.517490] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 581.517667] env[62000]: DEBUG oslo_concurrency.lockutils [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] Acquired lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.517705] env[62000]: DEBUG nova.network.neutron [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Refreshing network info cache for port 4b926553-f0d0-4888-a629-b4a735daa258 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 581.519115] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-df770652-fa19-4dfa-8da2-9a9ee7c33c53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.528933] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5311c8f0-b1ec-4c29-a030-1abd0e3b333c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.553265] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 57cc0a22-28ff-4a24-9fea-5c6a0a699a67 could not be found. [ 581.553541] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 581.553847] env[62000]: INFO nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Took 0.04 seconds to destroy the instance on the hypervisor. [ 581.553963] env[62000]: DEBUG oslo.service.loopingcall [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 581.554214] env[62000]: DEBUG nova.compute.manager [-] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 581.554308] env[62000]: DEBUG nova.network.neutron [-] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 581.599847] env[62000]: DEBUG nova.network.neutron [-] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 581.761989] env[62000]: DEBUG nova.network.neutron [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.817064] env[62000]: INFO nova.compute.manager [-] [instance: e99bfd97-1952-4166-aa19-560053bae307] Took 1.06 seconds to deallocate network for instance. [ 581.855657] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881833, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.903435] env[62000]: DEBUG nova.scheduler.client.report [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.106408] env[62000]: DEBUG nova.network.neutron [-] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.205475] env[62000]: DEBUG nova.network.neutron [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 582.265600] env[62000]: INFO nova.compute.manager [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Took 1.04 seconds to deallocate network for instance. [ 582.359162] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881833, 'name': ReconfigVM_Task, 'duration_secs': 0.782155} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.359162] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Reconfigured VM instance instance-00000009 to attach disk [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46/15409fa7-254c-435e-8080-46d3f65b2d46.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 582.359704] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a18634e0-8f3a-483b-8145-f6bcf11c3bcf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.362074] env[62000]: DEBUG nova.network.neutron [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 582.371604] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 582.371604] env[62000]: value = "task-881838" [ 582.371604] env[62000]: _type = "Task" [ 582.371604] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.380947] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881838, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.411185] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.411185] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.413953] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.782s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.552901] env[62000]: ERROR nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 582.552901] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.552901] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.552901] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.552901] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.552901] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.552901] env[62000]: ERROR nova.compute.manager raise self.value [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.552901] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.552901] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.552901] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.553903] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.553903] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.553903] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 582.553903] env[62000]: ERROR nova.compute.manager [ 582.553903] env[62000]: Traceback (most recent call last): [ 582.553903] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.553903] env[62000]: listener.cb(fileno) [ 582.553903] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.553903] env[62000]: result = function(*args, **kwargs) [ 582.553903] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.553903] env[62000]: return func(*args, **kwargs) [ 582.553903] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.553903] env[62000]: raise e [ 582.553903] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.553903] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 582.553903] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.553903] env[62000]: created_port_ids = self._update_ports_for_instance( [ 582.553903] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.553903] env[62000]: with excutils.save_and_reraise_exception(): [ 582.553903] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.553903] env[62000]: self.force_reraise() [ 582.553903] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.553903] env[62000]: raise self.value [ 582.553903] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.553903] env[62000]: updated_port = self._update_port( [ 582.553903] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.553903] env[62000]: _ensure_no_port_binding_failure(port) [ 582.553903] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.553903] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.554879] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 582.554879] env[62000]: Removing descriptor: 14 [ 582.554879] env[62000]: ERROR nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Traceback (most recent call last): [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] yield resources [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.driver.spawn(context, instance, image_meta, [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.554879] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] vm_ref = self.build_virtual_machine(instance, [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] for vif in network_info: [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self._sync_wrapper(fn, *args, **kwargs) [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.wait() [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self[:] = self._gt.wait() [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self._exit_event.wait() [ 582.555302] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] result = hub.switch() [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self.greenlet.switch() [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] result = function(*args, **kwargs) [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return func(*args, **kwargs) [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise e [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] nwinfo = self.network_api.allocate_for_instance( [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 582.555705] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] created_port_ids = self._update_ports_for_instance( [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] with excutils.save_and_reraise_exception(): [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.force_reraise() [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise self.value [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] updated_port = self._update_port( [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] _ensure_no_port_binding_failure(port) [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.556459] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise exception.PortBindingFailed(port_id=port['id']) [ 582.556870] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 582.556870] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] [ 582.556870] env[62000]: INFO nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Terminating instance [ 582.558835] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquiring lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.559035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquired lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.559192] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 582.612407] env[62000]: INFO nova.compute.manager [-] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Took 1.06 seconds to deallocate network for instance. [ 582.614777] env[62000]: DEBUG nova.compute.claims [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 582.614965] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.765820] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.766105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.857730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-21624707-f7f1-4188-b0d8-9d7f0fef1c94 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "e99bfd97-1952-4166-aa19-560053bae307" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.517s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.864507] env[62000]: DEBUG oslo_concurrency.lockutils [req-99be4e68-e14e-4365-b9e1-8443decf0072 req-18e4f807-7f85-43a3-a005-db1df64c75e6 service nova] Releasing lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.882263] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881838, 'name': Rename_Task, 'duration_secs': 0.15202} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.883492] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 582.883492] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c880ed39-3e2c-45fd-b3bb-dcd596df7416 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.891067] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Waiting for the task: (returnval){ [ 582.891067] env[62000]: value = "task-881839" [ 582.891067] env[62000]: _type = "Task" [ 582.891067] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.900072] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881839, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.919409] env[62000]: DEBUG nova.compute.utils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 582.924381] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 582.924553] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 582.992141] env[62000]: DEBUG nova.policy [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a8a4f278c38d4bcfb2be247f48ebc1ad', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b21cae77c3724d329b125a22e9d5f4a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.090760] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.178615] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.313449] env[62000]: INFO nova.scheduler.client.report [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Deleted allocations for instance fa76f279-9efd-401f-823e-806425ce39e5 [ 583.405199] env[62000]: DEBUG oslo_vmware.api [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Task: {'id': task-881839, 'name': PowerOnVM_Task, 'duration_secs': 0.487926} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.408087] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 583.408437] env[62000]: DEBUG nova.compute.manager [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 583.410355] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a129b0-a21d-4e2e-9ab1-2ec59532d0a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.416140] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.416511] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.424256] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ca0bb4-42d0-44bc-b22f-ee20d1f2915f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.427116] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.434695] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5ae067-af48-43bf-87a3-4eb2bba250c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.473525] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f907db6-d23c-49d3-9efa-f23b5750f684 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.476254] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Successfully created port: 329e734b-bbed-49c4-ab09-1a7b00390583 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 583.487801] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b521a7-3ad4-4e6a-9388-ae863d3561e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.499655] env[62000]: DEBUG nova.compute.provider_tree [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 583.503375] env[62000]: DEBUG nova.compute.manager [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Received event network-vif-deleted-4b926553-f0d0-4888-a629-b4a735daa258 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.503574] env[62000]: DEBUG nova.compute.manager [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Received event network-changed-16cacb86-1eef-4bc8-8bb8-19573360c52b {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 583.504091] env[62000]: DEBUG nova.compute.manager [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Refreshing instance network info cache due to event network-changed-16cacb86-1eef-4bc8-8bb8-19573360c52b. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 583.506939] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] Acquiring lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.686249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Releasing lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.686249] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.686249] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 583.686249] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] Acquired lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.686249] env[62000]: DEBUG nova.network.neutron [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Refreshing network info cache for port 16cacb86-1eef-4bc8-8bb8-19573360c52b {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 583.687497] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c2b7270c-bbc8-4821-b29b-c8e7305ceca8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.697774] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7f2b1e-6bbf-483f-ac9a-6da3dd100247 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.727397] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0f34d3e-39a1-4487-bd1f-ade41ef6832c could not be found. [ 583.727647] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 583.731148] env[62000]: INFO nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 583.731148] env[62000]: DEBUG oslo.service.loopingcall [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.731148] env[62000]: DEBUG nova.compute.manager [-] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.731148] env[62000]: DEBUG nova.network.neutron [-] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 583.756468] env[62000]: DEBUG nova.network.neutron [-] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 583.829156] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb88e77-d75d-4fed-8f19-b7f1c161b6e8 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.105s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.830700] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 14.462s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.830792] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "fa76f279-9efd-401f-823e-806425ce39e5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.831560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.831560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.834684] env[62000]: INFO nova.compute.manager [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Terminating instance [ 583.836724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquiring lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.836881] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Acquired lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.837063] env[62000]: DEBUG nova.network.neutron [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 583.942741] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.006323] env[62000]: DEBUG nova.scheduler.client.report [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.244123] env[62000]: DEBUG nova.network.neutron [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.259161] env[62000]: DEBUG nova.network.neutron [-] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.333505] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 584.379616] env[62000]: DEBUG nova.network.neutron [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.384077] env[62000]: DEBUG nova.network.neutron [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 584.441969] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.464730] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:53:18Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='20880505',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1762507564',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.464992] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.465168] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.465353] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.465504] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.465648] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.465849] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.466014] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.470853] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.470853] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.470853] env[62000]: DEBUG nova.virt.hardware [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.471986] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eceb04f2-c5b8-4f23-b002-99a9eb2cf147 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.486334] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ebda8d-00bf-4e54-ae72-1f950cae652a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.512136] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.098s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 584.512782] env[62000]: ERROR nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Traceback (most recent call last): [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.driver.spawn(context, instance, image_meta, [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] vm_ref = self.build_virtual_machine(instance, [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.512782] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] for vif in network_info: [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self._sync_wrapper(fn, *args, **kwargs) [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.wait() [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self[:] = self._gt.wait() [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self._exit_event.wait() [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] result = hub.switch() [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.513467] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return self.greenlet.switch() [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] result = function(*args, **kwargs) [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] return func(*args, **kwargs) [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise e [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] nwinfo = self.network_api.allocate_for_instance( [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] created_port_ids = self._update_ports_for_instance( [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] with excutils.save_and_reraise_exception(): [ 584.514039] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] self.force_reraise() [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise self.value [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] updated_port = self._update_port( [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] _ensure_no_port_binding_failure(port) [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] raise exception.PortBindingFailed(port_id=port['id']) [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] nova.exception.PortBindingFailed: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. [ 584.514766] env[62000]: ERROR nova.compute.manager [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] [ 584.515297] env[62000]: DEBUG nova.compute.utils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 584.515297] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.323s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.516444] env[62000]: INFO nova.compute.claims [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 584.519420] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Build of instance faa1794d-5941-40fb-bd78-2608fa2d5da0 was re-scheduled: Binding failed for port 3c71f5dc-6fa7-46bb-84cb-a3c384fb0585, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 584.519854] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 584.520090] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquiring lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.520743] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Acquired lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.520743] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 584.555998] env[62000]: DEBUG nova.network.neutron [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.761579] env[62000]: INFO nova.compute.manager [-] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Took 1.03 seconds to deallocate network for instance. [ 584.764390] env[62000]: DEBUG nova.compute.claims [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.764618] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.860219] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.882159] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ee9e3a3-57b6-4c5b-8165-e871f79953e1 req-4bd55ae8-e698-47a6-851b-a4b8ee511b41 service nova] Releasing lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.049292] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.059849] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Releasing lock "refresh_cache-fa76f279-9efd-401f-823e-806425ce39e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.060344] env[62000]: DEBUG nova.compute.manager [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.060560] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 585.060851] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c37d3996-8636-4df6-a5ae-dd64e12bcfa3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.069956] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc920c0-f8fc-4bce-92d5-9405d0c81b84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.094374] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa76f279-9efd-401f-823e-806425ce39e5 could not be found. [ 585.095023] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 585.095023] env[62000]: INFO nova.compute.manager [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 585.095339] env[62000]: DEBUG oslo.service.loopingcall [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.096319] env[62000]: ERROR nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 585.096319] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.096319] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.096319] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.096319] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.096319] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.096319] env[62000]: ERROR nova.compute.manager raise self.value [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.096319] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 585.096319] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.096319] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 585.097036] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.097036] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 585.097036] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 585.097036] env[62000]: ERROR nova.compute.manager [ 585.097036] env[62000]: Traceback (most recent call last): [ 585.097036] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 585.097036] env[62000]: listener.cb(fileno) [ 585.097036] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.097036] env[62000]: result = function(*args, **kwargs) [ 585.097036] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.097036] env[62000]: return func(*args, **kwargs) [ 585.097036] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.097036] env[62000]: raise e [ 585.097036] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.097036] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 585.097036] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.097036] env[62000]: created_port_ids = self._update_ports_for_instance( [ 585.097036] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.097036] env[62000]: with excutils.save_and_reraise_exception(): [ 585.097036] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.097036] env[62000]: self.force_reraise() [ 585.097036] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.097036] env[62000]: raise self.value [ 585.097036] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.097036] env[62000]: updated_port = self._update_port( [ 585.097036] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.097036] env[62000]: _ensure_no_port_binding_failure(port) [ 585.097036] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.097036] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 585.098032] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 585.098032] env[62000]: Removing descriptor: 14 [ 585.098032] env[62000]: DEBUG nova.compute.manager [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.098032] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.100562] env[62000]: ERROR nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Traceback (most recent call last): [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] yield resources [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.driver.spawn(context, instance, image_meta, [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] vm_ref = self.build_virtual_machine(instance, [ 585.100562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] for vif in network_info: [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self._sync_wrapper(fn, *args, **kwargs) [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.wait() [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self[:] = self._gt.wait() [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self._exit_event.wait() [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 585.101562] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] result = hub.switch() [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self.greenlet.switch() [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] result = function(*args, **kwargs) [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return func(*args, **kwargs) [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise e [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] nwinfo = self.network_api.allocate_for_instance( [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] created_port_ids = self._update_ports_for_instance( [ 585.104515] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] with excutils.save_and_reraise_exception(): [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.force_reraise() [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise self.value [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] updated_port = self._update_port( [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] _ensure_no_port_binding_failure(port) [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise exception.PortBindingFailed(port_id=port['id']) [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 585.104947] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] [ 585.105508] env[62000]: INFO nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Terminating instance [ 585.105508] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquiring lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.105508] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquired lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.105508] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.120464] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.184811] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.516596] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "15409fa7-254c-435e-8080-46d3f65b2d46" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.519776] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.519776] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "15409fa7-254c-435e-8080-46d3f65b2d46-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.519776] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.519776] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.520777] env[62000]: INFO nova.compute.manager [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Terminating instance [ 585.528557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "refresh_cache-15409fa7-254c-435e-8080-46d3f65b2d46" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.528557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquired lock "refresh_cache-15409fa7-254c-435e-8080-46d3f65b2d46" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 585.528557] env[62000]: DEBUG nova.network.neutron [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 585.624579] env[62000]: DEBUG nova.network.neutron [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.635562] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.689586] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Releasing lock "refresh_cache-faa1794d-5941-40fb-bd78-2608fa2d5da0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.689920] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 585.690098] env[62000]: DEBUG nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.690277] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 585.711506] env[62000]: DEBUG nova.compute.manager [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Received event network-vif-deleted-16cacb86-1eef-4bc8-8bb8-19573360c52b {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 585.711506] env[62000]: DEBUG nova.compute.manager [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Received event network-changed-329e734b-bbed-49c4-ab09-1a7b00390583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 585.711639] env[62000]: DEBUG nova.compute.manager [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Refreshing instance network info cache due to event network-changed-329e734b-bbed-49c4-ab09-1a7b00390583. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 585.712070] env[62000]: DEBUG oslo_concurrency.lockutils [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] Acquiring lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 585.737339] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 585.880395] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.039500] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0d9b511-2a1e-47f3-b5d2-9478c99a7796 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.047548] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2483bde8-10a8-4901-b323-e42ada98b0c6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.084927] env[62000]: DEBUG nova.network.neutron [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.088033] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b89acee-02f5-47c0-9ce4-5f3ad39d0382 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.096418] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea70cab-2eb8-4443-bd7f-975d303274c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.112623] env[62000]: DEBUG nova.compute.provider_tree [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.128924] env[62000]: INFO nova.compute.manager [-] [instance: fa76f279-9efd-401f-823e-806425ce39e5] Took 1.03 seconds to deallocate network for instance. [ 586.244745] env[62000]: DEBUG nova.network.neutron [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.265178] env[62000]: DEBUG nova.network.neutron [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.384469] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Releasing lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.384469] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.384469] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 586.384716] env[62000]: DEBUG oslo_concurrency.lockutils [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] Acquired lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.384716] env[62000]: DEBUG nova.network.neutron [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Refreshing network info cache for port 329e734b-bbed-49c4-ab09-1a7b00390583 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 586.386935] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e038f02-5e64-4665-a0df-df8f09d220d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.397906] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bbf3f95-2d53-48e4-99a9-8ee8f0956efb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.426964] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 465d94c2-5ebf-4b05-966e-9272e2ddebb3 could not be found. [ 586.427239] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 586.427435] env[62000]: INFO nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 586.427682] env[62000]: DEBUG oslo.service.loopingcall [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 586.427907] env[62000]: DEBUG nova.compute.manager [-] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 586.428009] env[62000]: DEBUG nova.network.neutron [-] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 586.501140] env[62000]: DEBUG nova.network.neutron [-] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 586.616091] env[62000]: DEBUG nova.scheduler.client.report [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.745265] env[62000]: INFO nova.compute.manager [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] [instance: faa1794d-5941-40fb-bd78-2608fa2d5da0] Took 1.05 seconds to deallocate network for instance. [ 586.771088] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Releasing lock "refresh_cache-15409fa7-254c-435e-8080-46d3f65b2d46" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 586.772037] env[62000]: DEBUG nova.compute.manager [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 586.772037] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 586.772698] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c180af8f-0866-4c02-a654-ae6bde21e9cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.785632] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 586.785632] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7cb19051-dd58-4243-be18-98341a0a1048 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.790740] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 586.790740] env[62000]: value = "task-881841" [ 586.790740] env[62000]: _type = "Task" [ 586.790740] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 586.799769] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881841, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 586.926888] env[62000]: DEBUG nova.network.neutron [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 587.004186] env[62000]: DEBUG nova.network.neutron [-] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.120956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.121528] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.125533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.886s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.127453] env[62000]: INFO nova.compute.claims [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.157568] env[62000]: DEBUG nova.network.neutron [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.166305] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19dea122-c261-464f-a5ab-a8f0e1b60d18 tempest-ListServersNegativeTestJSON-1740241068 tempest-ListServersNegativeTestJSON-1740241068-project-member] Lock "fa76f279-9efd-401f-823e-806425ce39e5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.335s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.302980] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881841, 'name': PowerOffVM_Task, 'duration_secs': 0.166082} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.302980] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 587.303166] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 587.303409] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a452ce86-fd39-489d-951e-9ffd5c79c481 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.329668] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 587.329904] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 587.330099] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleting the datastore file [datastore1] 15409fa7-254c-435e-8080-46d3f65b2d46 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 587.330352] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c4794609-ce97-414a-924f-726f282b8342 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.338766] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for the task: (returnval){ [ 587.338766] env[62000]: value = "task-881844" [ 587.338766] env[62000]: _type = "Task" [ 587.338766] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 587.346805] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881844, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 587.510574] env[62000]: INFO nova.compute.manager [-] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Took 1.08 seconds to deallocate network for instance. [ 587.513312] env[62000]: DEBUG nova.compute.claims [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 587.513501] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.634476] env[62000]: DEBUG nova.compute.utils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 587.643594] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 587.643783] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 587.659621] env[62000]: DEBUG oslo_concurrency.lockutils [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] Releasing lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.660090] env[62000]: DEBUG nova.compute.manager [req-ca19cfe5-5c57-4ef6-a7ad-ab4444bdb844 req-a636ed2b-d518-4101-adff-a5fd910f8529 service nova] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Received event network-vif-deleted-329e734b-bbed-49c4-ab09-1a7b00390583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 587.751917] env[62000]: DEBUG nova.policy [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fabd6e8940941ada9540441560393a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad12e169457d491097ab311e05d8e30a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 587.785688] env[62000]: INFO nova.scheduler.client.report [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Deleted allocations for instance faa1794d-5941-40fb-bd78-2608fa2d5da0 [ 587.850707] env[62000]: DEBUG oslo_vmware.api [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Task: {'id': task-881844, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.097861} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 587.851354] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 587.852467] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 587.852467] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 587.852467] env[62000]: INFO nova.compute.manager [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Took 1.08 seconds to destroy the instance on the hypervisor. [ 587.852467] env[62000]: DEBUG oslo.service.loopingcall [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.853248] env[62000]: DEBUG nova.compute.manager [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.853248] env[62000]: DEBUG nova.network.neutron [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 587.888096] env[62000]: DEBUG nova.network.neutron [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 588.147782] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.254356] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Successfully created port: 791a6040-4ad9-4150-8630-48761277b78d {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 588.300153] env[62000]: DEBUG oslo_concurrency.lockutils [None req-275f4d82-7ddf-461d-b854-5ca4e376a6ef tempest-ServersTestJSON-275757668 tempest-ServersTestJSON-275757668-project-member] Lock "faa1794d-5941-40fb-bd78-2608fa2d5da0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.839s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 588.388994] env[62000]: DEBUG nova.network.neutron [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.653863] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-449fdb67-6c6f-4fd5-8a67-7d702c65e98d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.665544] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da19e096-0864-41ae-8e73-dc074c4241c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.704333] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b37b505-c5d6-47a0-b1d8-fbe6e379399d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.713903] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc294368-f26a-4313-be07-96a61f1ead5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.735777] env[62000]: DEBUG nova.compute.provider_tree [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.803864] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.892672] env[62000]: INFO nova.compute.manager [-] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Took 1.04 seconds to deallocate network for instance. [ 589.165351] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.199260] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.199521] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.199677] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.199867] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.200353] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.200582] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.200828] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.201094] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.201272] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.201440] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.201626] env[62000]: DEBUG nova.virt.hardware [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.202586] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f330f30d-e0ea-457a-9a52-6625b9c72ffb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.213454] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-256cfe23-040a-45d4-8e10-c2b05cd0584d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.239842] env[62000]: DEBUG nova.scheduler.client.report [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.331687] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.404511] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.747451] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.747985] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.752242] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.490s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.756661] env[62000]: INFO nova.compute.claims [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 589.779394] env[62000]: ERROR nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 589.779394] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.779394] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.779394] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.779394] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.779394] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.779394] env[62000]: ERROR nova.compute.manager raise self.value [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.779394] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.779394] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.779394] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.779962] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.779962] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.779962] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 589.779962] env[62000]: ERROR nova.compute.manager [ 589.779962] env[62000]: Traceback (most recent call last): [ 589.779962] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.779962] env[62000]: listener.cb(fileno) [ 589.779962] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.779962] env[62000]: result = function(*args, **kwargs) [ 589.779962] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.779962] env[62000]: return func(*args, **kwargs) [ 589.779962] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.779962] env[62000]: raise e [ 589.779962] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.779962] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 589.779962] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.779962] env[62000]: created_port_ids = self._update_ports_for_instance( [ 589.779962] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.779962] env[62000]: with excutils.save_and_reraise_exception(): [ 589.779962] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.779962] env[62000]: self.force_reraise() [ 589.779962] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.779962] env[62000]: raise self.value [ 589.779962] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.779962] env[62000]: updated_port = self._update_port( [ 589.779962] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.779962] env[62000]: _ensure_no_port_binding_failure(port) [ 589.779962] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.779962] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.780953] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 589.780953] env[62000]: Removing descriptor: 16 [ 589.780953] env[62000]: ERROR nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] Traceback (most recent call last): [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] yield resources [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.driver.spawn(context, instance, image_meta, [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.780953] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] vm_ref = self.build_virtual_machine(instance, [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] for vif in network_info: [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self._sync_wrapper(fn, *args, **kwargs) [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.wait() [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self[:] = self._gt.wait() [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self._exit_event.wait() [ 589.781366] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] result = hub.switch() [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self.greenlet.switch() [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] result = function(*args, **kwargs) [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return func(*args, **kwargs) [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise e [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] nwinfo = self.network_api.allocate_for_instance( [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 589.781821] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] created_port_ids = self._update_ports_for_instance( [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] with excutils.save_and_reraise_exception(): [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.force_reraise() [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise self.value [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] updated_port = self._update_port( [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] _ensure_no_port_binding_failure(port) [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.782302] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise exception.PortBindingFailed(port_id=port['id']) [ 589.782723] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 589.782723] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] [ 589.782723] env[62000]: INFO nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Terminating instance [ 589.782723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.782873] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.783573] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 589.825512] env[62000]: DEBUG nova.compute.manager [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Received event network-changed-791a6040-4ad9-4150-8630-48761277b78d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 589.825724] env[62000]: DEBUG nova.compute.manager [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Refreshing instance network info cache due to event network-changed-791a6040-4ad9-4150-8630-48761277b78d. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 589.825914] env[62000]: DEBUG oslo_concurrency.lockutils [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] Acquiring lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.264205] env[62000]: DEBUG nova.compute.utils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 590.274361] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 590.274361] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 590.328682] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 590.432157] env[62000]: DEBUG nova.policy [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '187d789b014a4e6ba7fdb719f1ad0db0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9f40435bd544efcada1427af9fc8596', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.593747] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.773249] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 591.096988] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.097452] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 591.097640] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 591.100998] env[62000]: DEBUG oslo_concurrency.lockutils [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] Acquired lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.100998] env[62000]: DEBUG nova.network.neutron [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Refreshing network info cache for port 791a6040-4ad9-4150-8630-48761277b78d {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 591.102010] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a14ee0f1-9759-48f2-adb6-924f2777b57c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.112208] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498138dd-6b96-45a1-8c39-6639dc2a13c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.131924] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Successfully created port: 09c8357e-8463-4f6f-8946-da3c798950eb {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.152162] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d7c25703-df04-4289-92d6-226180c43082 could not be found. [ 591.152347] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 591.152788] env[62000]: INFO nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Took 0.05 seconds to destroy the instance on the hypervisor. [ 591.153300] env[62000]: DEBUG oslo.service.loopingcall [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 591.155661] env[62000]: DEBUG nova.compute.manager [-] [instance: d7c25703-df04-4289-92d6-226180c43082] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.155831] env[62000]: DEBUG nova.network.neutron [-] [instance: d7c25703-df04-4289-92d6-226180c43082] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 591.190492] env[62000]: DEBUG nova.network.neutron [-] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.328442] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6b3b803-7f52-4947-bc79-d7499e667cd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.335654] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d359785-07dd-444c-98e0-f8849c8e3c6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.377253] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918c2734-7ae2-4f68-9a1d-956644aa7b92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.386618] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715805f4-ea06-4d9b-8117-940f9f038fe1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.407248] env[62000]: DEBUG nova.compute.provider_tree [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 591.648049] env[62000]: DEBUG nova.network.neutron [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 591.695468] env[62000]: DEBUG nova.network.neutron [-] [instance: d7c25703-df04-4289-92d6-226180c43082] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.793024] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.827975] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.828300] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.828500] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.828722] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.828910] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.829356] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.831025] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.833182] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.833182] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.833182] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.833182] env[62000]: DEBUG nova.virt.hardware [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.833182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609373d1-9366-4d53-87c8-af6179ca8856 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.841601] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cfe5053-22d6-4f09-8691-5c61fb793e78 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.910452] env[62000]: DEBUG nova.scheduler.client.report [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 592.034882] env[62000]: DEBUG nova.network.neutron [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.200760] env[62000]: INFO nova.compute.manager [-] [instance: d7c25703-df04-4289-92d6-226180c43082] Took 1.04 seconds to deallocate network for instance. [ 592.203735] env[62000]: DEBUG nova.compute.claims [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.203735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.239497] env[62000]: DEBUG nova.compute.manager [req-38c921bd-1baa-4e9b-beb2-4ba14250996d req-56264d0c-fb87-4ac0-b720-8bc3553147dc service nova] [instance: d7c25703-df04-4289-92d6-226180c43082] Received event network-vif-deleted-791a6040-4ad9-4150-8630-48761277b78d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 592.416031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.662s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 592.416031] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 592.417972] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.362s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.423554] env[62000]: INFO nova.compute.claims [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 592.537790] env[62000]: DEBUG oslo_concurrency.lockutils [req-4828d5ad-d590-4349-a313-3885e0b26b0a req-169e5d7d-217b-4714-a1c8-0b2e13c98c71 service nova] Releasing lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.929166] env[62000]: DEBUG nova.compute.utils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 592.931203] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 592.931377] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 593.006250] env[62000]: DEBUG nova.policy [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '78f14539add54eb5abb18bfb664a970b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b7cdd441e524beba598b4d490e27592', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 593.432733] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 593.762332] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Successfully created port: c7a1e502-23ab-4f8b-81cc-95f8b94dbd79 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 593.989167] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f23105-3062-418a-a039-064a06650f56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.000622] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f10cbab-9f5e-48a5-8a94-272813f92884 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.039016] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07bd29f0-ccca-4afd-9d61-aae6309bd707 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.047889] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5feacec4-7d8c-49f4-a1e1-49d0050d0c0f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.064220] env[62000]: DEBUG nova.compute.provider_tree [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 594.084682] env[62000]: ERROR nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 594.084682] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.084682] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.084682] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.084682] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.084682] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.084682] env[62000]: ERROR nova.compute.manager raise self.value [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.084682] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.084682] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.084682] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.085291] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.085291] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.085291] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 594.085291] env[62000]: ERROR nova.compute.manager [ 594.085291] env[62000]: Traceback (most recent call last): [ 594.085291] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.085291] env[62000]: listener.cb(fileno) [ 594.085291] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.085291] env[62000]: result = function(*args, **kwargs) [ 594.085291] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.085291] env[62000]: return func(*args, **kwargs) [ 594.085291] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.085291] env[62000]: raise e [ 594.085291] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.085291] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 594.085291] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.085291] env[62000]: created_port_ids = self._update_ports_for_instance( [ 594.085291] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.085291] env[62000]: with excutils.save_and_reraise_exception(): [ 594.085291] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.085291] env[62000]: self.force_reraise() [ 594.085291] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.085291] env[62000]: raise self.value [ 594.085291] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.085291] env[62000]: updated_port = self._update_port( [ 594.085291] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.085291] env[62000]: _ensure_no_port_binding_failure(port) [ 594.085291] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.085291] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.086300] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 594.086300] env[62000]: Removing descriptor: 16 [ 594.086300] env[62000]: ERROR nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Traceback (most recent call last): [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] yield resources [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.driver.spawn(context, instance, image_meta, [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.086300] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] vm_ref = self.build_virtual_machine(instance, [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] for vif in network_info: [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self._sync_wrapper(fn, *args, **kwargs) [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.wait() [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self[:] = self._gt.wait() [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self._exit_event.wait() [ 594.086669] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] result = hub.switch() [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self.greenlet.switch() [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] result = function(*args, **kwargs) [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return func(*args, **kwargs) [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise e [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] nwinfo = self.network_api.allocate_for_instance( [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 594.087076] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] created_port_ids = self._update_ports_for_instance( [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] with excutils.save_and_reraise_exception(): [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.force_reraise() [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise self.value [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] updated_port = self._update_port( [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] _ensure_no_port_binding_failure(port) [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.087453] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise exception.PortBindingFailed(port_id=port['id']) [ 594.087818] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 594.087818] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] [ 594.087818] env[62000]: INFO nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Terminating instance [ 594.092631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.092631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquired lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.092631] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 594.457498] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 594.504491] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 594.504775] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 594.504900] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 594.506972] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 594.507209] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 594.507341] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 594.507532] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 594.507729] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 594.507856] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 594.508127] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 594.508207] env[62000]: DEBUG nova.virt.hardware [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 594.509064] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282af367-4100-4d6d-b2ba-0f3f4a7ce0a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.522658] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa5463d8-fb55-4025-8a10-4152cd59d0cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.569596] env[62000]: DEBUG nova.scheduler.client.report [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 594.628974] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 594.644355] env[62000]: DEBUG nova.compute.manager [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Received event network-changed-09c8357e-8463-4f6f-8946-da3c798950eb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 594.646487] env[62000]: DEBUG nova.compute.manager [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Refreshing instance network info cache due to event network-changed-09c8357e-8463-4f6f-8946-da3c798950eb. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 594.646487] env[62000]: DEBUG oslo_concurrency.lockutils [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] Acquiring lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.764038] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.083360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.659s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.083360] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 595.086281] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 22.885s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.086281] env[62000]: DEBUG nova.objects.instance [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 595.244995] env[62000]: ERROR nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 595.244995] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.244995] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.244995] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.244995] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.244995] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.244995] env[62000]: ERROR nova.compute.manager raise self.value [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.244995] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 595.244995] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.244995] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 595.245521] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.245521] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 595.245521] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 595.245521] env[62000]: ERROR nova.compute.manager [ 595.245521] env[62000]: Traceback (most recent call last): [ 595.245521] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 595.245521] env[62000]: listener.cb(fileno) [ 595.245521] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.245521] env[62000]: result = function(*args, **kwargs) [ 595.245521] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.245521] env[62000]: return func(*args, **kwargs) [ 595.245521] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.245521] env[62000]: raise e [ 595.245521] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.245521] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 595.245521] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.245521] env[62000]: created_port_ids = self._update_ports_for_instance( [ 595.245521] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.245521] env[62000]: with excutils.save_and_reraise_exception(): [ 595.245521] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.245521] env[62000]: self.force_reraise() [ 595.245521] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.245521] env[62000]: raise self.value [ 595.245521] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.245521] env[62000]: updated_port = self._update_port( [ 595.245521] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.245521] env[62000]: _ensure_no_port_binding_failure(port) [ 595.245521] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.245521] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 595.246345] env[62000]: nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 595.246345] env[62000]: Removing descriptor: 14 [ 595.246765] env[62000]: ERROR nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Traceback (most recent call last): [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] yield resources [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.driver.spawn(context, instance, image_meta, [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] vm_ref = self.build_virtual_machine(instance, [ 595.246765] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] vif_infos = vmwarevif.get_vif_info(self._session, [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] for vif in network_info: [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self._sync_wrapper(fn, *args, **kwargs) [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.wait() [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self[:] = self._gt.wait() [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self._exit_event.wait() [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 595.247280] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] result = hub.switch() [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self.greenlet.switch() [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] result = function(*args, **kwargs) [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return func(*args, **kwargs) [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise e [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] nwinfo = self.network_api.allocate_for_instance( [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] created_port_ids = self._update_ports_for_instance( [ 595.249716] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] with excutils.save_and_reraise_exception(): [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.force_reraise() [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise self.value [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] updated_port = self._update_port( [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] _ensure_no_port_binding_failure(port) [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise exception.PortBindingFailed(port_id=port['id']) [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 595.250226] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] [ 595.250671] env[62000]: INFO nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Terminating instance [ 595.250671] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquiring lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.250671] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquired lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.250671] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 595.266085] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Releasing lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.266964] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.266964] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 595.267729] env[62000]: DEBUG oslo_concurrency.lockutils [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] Acquired lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.267729] env[62000]: DEBUG nova.network.neutron [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Refreshing network info cache for port 09c8357e-8463-4f6f-8946-da3c798950eb {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 595.268411] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41bd7c7c-5c07-4a37-b1f9-ed8e33fbb28e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.278644] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3b4467-8444-4d58-87eb-f73bb4d44b7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.302835] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a36de4b6-6928-4110-bc27-825ae58b15e5 could not be found. [ 595.303104] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 595.303294] env[62000]: INFO nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 595.303539] env[62000]: DEBUG oslo.service.loopingcall [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.303753] env[62000]: DEBUG nova.compute.manager [-] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.303840] env[62000]: DEBUG nova.network.neutron [-] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 595.330743] env[62000]: DEBUG nova.network.neutron [-] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.397488] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 595.397488] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 595.590271] env[62000]: DEBUG nova.compute.utils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 595.596741] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 595.596933] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 595.690103] env[62000]: DEBUG nova.policy [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a383ddc61034bf5b48be4a55ad9c330', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9de9f8fa96fd42ea9f9ba9698a74228d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 595.782032] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.799858] env[62000]: DEBUG nova.network.neutron [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 595.833324] env[62000]: DEBUG nova.network.neutron [-] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.000518] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.021202] env[62000]: DEBUG nova.network.neutron [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.099715] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 596.103023] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d39cd96-f88d-4300-8176-9e3a258b5461 tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.106102] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.720s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.106102] env[62000]: INFO nova.compute.claims [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 596.235309] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.235559] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.336792] env[62000]: INFO nova.compute.manager [-] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Took 1.03 seconds to deallocate network for instance. [ 596.340370] env[62000]: DEBUG nova.compute.claims [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.340691] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.343527] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Successfully created port: ebc7822a-dc8c-477e-8899-d913cd24e66a {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 596.509034] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Releasing lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.509034] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 596.509034] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 596.509034] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b6d0c448-94c9-4f49-b188-4b43af77f789 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.516944] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd04160e-1e1f-47b2-8e07-4d4a714e983e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.529829] env[62000]: DEBUG oslo_concurrency.lockutils [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] Releasing lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.529975] env[62000]: DEBUG nova.compute.manager [req-a59cf0a8-5f18-4be3-9890-2971d4135b80 req-1de8446e-dd1e-49bf-84f0-85a07a86fb35 service nova] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Received event network-vif-deleted-09c8357e-8463-4f6f-8946-da3c798950eb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 596.545526] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f could not be found. [ 596.547377] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 596.547377] env[62000]: INFO nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 596.547377] env[62000]: DEBUG oslo.service.loopingcall [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 596.547377] env[62000]: DEBUG nova.compute.manager [-] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 596.547377] env[62000]: DEBUG nova.network.neutron [-] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 596.591331] env[62000]: DEBUG nova.network.neutron [-] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.092095] env[62000]: DEBUG nova.network.neutron [-] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.111661] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 597.155184] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 597.155393] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 597.155557] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 597.155758] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 597.157058] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 597.157317] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 597.157856] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 597.157856] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 597.157856] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 597.158640] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 597.158640] env[62000]: DEBUG nova.virt.hardware [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 597.159485] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022a8b74-f4d3-4f0f-ba24-31bd7609c7d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.175524] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d7818f-a85b-48fd-b435-8294f2f22072 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.201461] env[62000]: DEBUG nova.compute.manager [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Received event network-changed-c7a1e502-23ab-4f8b-81cc-95f8b94dbd79 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 597.201655] env[62000]: DEBUG nova.compute.manager [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Refreshing instance network info cache due to event network-changed-c7a1e502-23ab-4f8b-81cc-95f8b94dbd79. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 597.201934] env[62000]: DEBUG oslo_concurrency.lockutils [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] Acquiring lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.202097] env[62000]: DEBUG oslo_concurrency.lockutils [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] Acquired lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 597.202259] env[62000]: DEBUG nova.network.neutron [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Refreshing network info cache for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 597.220754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "72688992-333f-459d-9d05-f7c728961a6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.220754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.594886] env[62000]: INFO nova.compute.manager [-] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Took 1.05 seconds to deallocate network for instance. [ 597.600516] env[62000]: DEBUG nova.compute.claims [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 597.600796] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.675574] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9916560-6850-4529-92f7-415526eb9d3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.684864] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0baf8e-1bfd-4009-ad1e-0e3e5c84e6ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.718628] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac197f52-f32e-4e28-942d-3715e354b8c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.726275] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd593607-57c5-417c-902d-37c878eaaff7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.740807] env[62000]: DEBUG nova.compute.provider_tree [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.749108] env[62000]: DEBUG nova.network.neutron [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 597.904975] env[62000]: DEBUG nova.network.neutron [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.244566] env[62000]: DEBUG nova.scheduler.client.report [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.285945] env[62000]: ERROR nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 598.285945] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.285945] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.285945] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.285945] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.285945] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.285945] env[62000]: ERROR nova.compute.manager raise self.value [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.285945] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 598.285945] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.285945] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 598.286428] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.286428] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 598.286428] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 598.286428] env[62000]: ERROR nova.compute.manager [ 598.286428] env[62000]: Traceback (most recent call last): [ 598.286428] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 598.286428] env[62000]: listener.cb(fileno) [ 598.286428] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.286428] env[62000]: result = function(*args, **kwargs) [ 598.286428] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.286428] env[62000]: return func(*args, **kwargs) [ 598.286428] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.286428] env[62000]: raise e [ 598.286428] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.286428] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 598.286428] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.286428] env[62000]: created_port_ids = self._update_ports_for_instance( [ 598.286428] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.286428] env[62000]: with excutils.save_and_reraise_exception(): [ 598.286428] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.286428] env[62000]: self.force_reraise() [ 598.286428] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.286428] env[62000]: raise self.value [ 598.286428] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.286428] env[62000]: updated_port = self._update_port( [ 598.286428] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.286428] env[62000]: _ensure_no_port_binding_failure(port) [ 598.286428] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.286428] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 598.287773] env[62000]: nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 598.287773] env[62000]: Removing descriptor: 14 [ 598.287773] env[62000]: ERROR nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Traceback (most recent call last): [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] yield resources [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.driver.spawn(context, instance, image_meta, [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.287773] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] vm_ref = self.build_virtual_machine(instance, [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] for vif in network_info: [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self._sync_wrapper(fn, *args, **kwargs) [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.wait() [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self[:] = self._gt.wait() [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self._exit_event.wait() [ 598.288151] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] result = hub.switch() [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self.greenlet.switch() [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] result = function(*args, **kwargs) [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return func(*args, **kwargs) [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise e [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] nwinfo = self.network_api.allocate_for_instance( [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 598.288606] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] created_port_ids = self._update_ports_for_instance( [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] with excutils.save_and_reraise_exception(): [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.force_reraise() [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise self.value [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] updated_port = self._update_port( [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] _ensure_no_port_binding_failure(port) [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.288987] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise exception.PortBindingFailed(port_id=port['id']) [ 598.289365] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 598.289365] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] [ 598.289365] env[62000]: INFO nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Terminating instance [ 598.293081] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquiring lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.293081] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquired lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.293081] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 598.409910] env[62000]: DEBUG oslo_concurrency.lockutils [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] Releasing lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.409910] env[62000]: DEBUG nova.compute.manager [req-4aa46951-21bd-4794-bc74-45b85623a013 req-5dd21b01-79e1-4fc0-a0aa-49732eee7024 service nova] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Received event network-vif-deleted-c7a1e502-23ab-4f8b-81cc-95f8b94dbd79 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 598.750745] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.647s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.751309] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 598.759408] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.141s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.820701] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 598.999890] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.206770] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.207042] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.230171] env[62000]: DEBUG nova.compute.manager [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Received event network-changed-ebc7822a-dc8c-477e-8899-d913cd24e66a {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 599.230390] env[62000]: DEBUG nova.compute.manager [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Refreshing instance network info cache due to event network-changed-ebc7822a-dc8c-477e-8899-d913cd24e66a. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 599.230573] env[62000]: DEBUG oslo_concurrency.lockutils [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] Acquiring lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 599.258355] env[62000]: DEBUG nova.compute.utils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 599.260060] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 599.260060] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 599.369749] env[62000]: DEBUG nova.policy [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ab2618736624ba38218e9f63cf584e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4aac407ee1974bc295ee88e99d651611', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 599.504963] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Releasing lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.505476] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 599.505635] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 599.505954] env[62000]: DEBUG oslo_concurrency.lockutils [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] Acquired lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 599.506142] env[62000]: DEBUG nova.network.neutron [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Refreshing network info cache for port ebc7822a-dc8c-477e-8899-d913cd24e66a {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 599.507596] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-666b6f55-942b-47ff-9fdc-ea7605534ee0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.520362] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f80de36-77ee-471a-aaa5-554490730736 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.546646] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3bd34ec2-5253-4578-baf3-11775de18ed4 could not be found. [ 599.546880] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 599.549191] env[62000]: INFO nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 599.549191] env[62000]: DEBUG oslo.service.loopingcall [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 599.549879] env[62000]: DEBUG nova.compute.manager [-] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.550063] env[62000]: DEBUG nova.network.neutron [-] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 599.591508] env[62000]: DEBUG nova.network.neutron [-] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 599.765050] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 599.789195] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad69b13a-0de8-4837-87c0-c0e578a2f19b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.797417] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3f906e-fb9d-49ff-a55b-3310c66daf4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.837447] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07c3dd81-e99e-429e-8f61-ff8c0a7163ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.846430] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1029f77-d959-47c5-80f5-98b835a39f9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.862268] env[62000]: DEBUG nova.compute.provider_tree [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.937562] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Successfully created port: a8099270-f889-49e5-8e92-13e439ea9de1 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 600.042560] env[62000]: DEBUG nova.network.neutron [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 600.093777] env[62000]: DEBUG nova.network.neutron [-] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.281402] env[62000]: DEBUG nova.network.neutron [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.363288] env[62000]: DEBUG nova.scheduler.client.report [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.598826] env[62000]: INFO nova.compute.manager [-] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Took 1.05 seconds to deallocate network for instance. [ 600.605385] env[62000]: DEBUG nova.compute.claims [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 600.605628] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.773812] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 600.784291] env[62000]: DEBUG oslo_concurrency.lockutils [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] Releasing lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 600.784608] env[62000]: DEBUG nova.compute.manager [req-f073a2b9-7285-4845-9098-78c6a5be88a5 req-232a1bf1-0cc0-41ba-bf4e-fcaa39575645 service nova] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Received event network-vif-deleted-ebc7822a-dc8c-477e-8899-d913cd24e66a {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 600.811538] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 600.811787] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 600.812011] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 600.812232] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 600.812372] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 600.812518] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 600.812834] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 600.813563] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 600.813825] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 600.814014] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 600.814237] env[62000]: DEBUG nova.virt.hardware [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 600.817232] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63951b5f-b938-4b45-a912-e752f0fe470c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.829977] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4840bc08-f772-4ec7-966f-672f399fa2d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.868766] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.112s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.869424] env[62000]: ERROR nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Traceback (most recent call last): [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.driver.spawn(context, instance, image_meta, [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] vm_ref = self.build_virtual_machine(instance, [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.869424] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] for vif in network_info: [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self._sync_wrapper(fn, *args, **kwargs) [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.wait() [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self[:] = self._gt.wait() [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self._exit_event.wait() [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] result = hub.switch() [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.869808] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return self.greenlet.switch() [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] result = function(*args, **kwargs) [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] return func(*args, **kwargs) [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise e [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] nwinfo = self.network_api.allocate_for_instance( [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] created_port_ids = self._update_ports_for_instance( [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] with excutils.save_and_reraise_exception(): [ 600.870236] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] self.force_reraise() [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise self.value [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] updated_port = self._update_port( [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] _ensure_no_port_binding_failure(port) [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] raise exception.PortBindingFailed(port_id=port['id']) [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] nova.exception.PortBindingFailed: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. [ 600.870632] env[62000]: ERROR nova.compute.manager [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] [ 600.870961] env[62000]: DEBUG nova.compute.utils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 600.871505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.929s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.871633] env[62000]: DEBUG nova.objects.instance [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 600.874424] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Build of instance 57cc0a22-28ff-4a24-9fea-5c6a0a699a67 was re-scheduled: Binding failed for port 4b926553-f0d0-4888-a629-b4a735daa258, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 600.874881] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 600.875119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquiring lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.875269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Acquired lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.875437] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 601.137441] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 601.137673] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.283925] env[62000]: DEBUG nova.compute.manager [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Received event network-changed-a8099270-f889-49e5-8e92-13e439ea9de1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 601.284395] env[62000]: DEBUG nova.compute.manager [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Refreshing instance network info cache due to event network-changed-a8099270-f889-49e5-8e92-13e439ea9de1. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 601.284606] env[62000]: DEBUG oslo_concurrency.lockutils [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] Acquiring lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.284810] env[62000]: DEBUG oslo_concurrency.lockutils [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] Acquired lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.284908] env[62000]: DEBUG nova.network.neutron [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Refreshing network info cache for port a8099270-f889-49e5-8e92-13e439ea9de1 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 601.401574] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.482302] env[62000]: ERROR nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 601.482302] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.482302] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.482302] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.482302] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.482302] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.482302] env[62000]: ERROR nova.compute.manager raise self.value [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.482302] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 601.482302] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.482302] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 601.483643] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.483643] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 601.483643] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 601.483643] env[62000]: ERROR nova.compute.manager [ 601.483643] env[62000]: Traceback (most recent call last): [ 601.483643] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 601.483643] env[62000]: listener.cb(fileno) [ 601.483643] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.483643] env[62000]: result = function(*args, **kwargs) [ 601.483643] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.483643] env[62000]: return func(*args, **kwargs) [ 601.483643] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.483643] env[62000]: raise e [ 601.483643] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.483643] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 601.483643] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.483643] env[62000]: created_port_ids = self._update_ports_for_instance( [ 601.483643] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.483643] env[62000]: with excutils.save_and_reraise_exception(): [ 601.483643] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.483643] env[62000]: self.force_reraise() [ 601.483643] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.483643] env[62000]: raise self.value [ 601.483643] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.483643] env[62000]: updated_port = self._update_port( [ 601.483643] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.483643] env[62000]: _ensure_no_port_binding_failure(port) [ 601.483643] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.483643] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 601.485053] env[62000]: nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 601.485053] env[62000]: Removing descriptor: 14 [ 601.485053] env[62000]: ERROR nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Traceback (most recent call last): [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] yield resources [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.driver.spawn(context, instance, image_meta, [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.485053] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] vm_ref = self.build_virtual_machine(instance, [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] for vif in network_info: [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self._sync_wrapper(fn, *args, **kwargs) [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.wait() [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self[:] = self._gt.wait() [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self._exit_event.wait() [ 601.485613] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] result = hub.switch() [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self.greenlet.switch() [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] result = function(*args, **kwargs) [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return func(*args, **kwargs) [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise e [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] nwinfo = self.network_api.allocate_for_instance( [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 601.486140] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] created_port_ids = self._update_ports_for_instance( [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] with excutils.save_and_reraise_exception(): [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.force_reraise() [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise self.value [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] updated_port = self._update_port( [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] _ensure_no_port_binding_failure(port) [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.486516] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise exception.PortBindingFailed(port_id=port['id']) [ 601.487204] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 601.487204] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] [ 601.487204] env[62000]: INFO nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Terminating instance [ 601.488791] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquiring lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.523284] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.817253] env[62000]: DEBUG nova.network.neutron [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 601.888225] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd90b1dd-a404-4a90-8494-1004758aa26e tempest-ServersAdmin275Test-2003862271 tempest-ServersAdmin275Test-2003862271-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.890698] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.126s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.916215] env[62000]: DEBUG nova.network.neutron [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.025889] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Releasing lock "refresh_cache-57cc0a22-28ff-4a24-9fea-5c6a0a699a67" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.027040] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 602.027040] env[62000]: DEBUG nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.027040] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 602.043449] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 602.418609] env[62000]: DEBUG oslo_concurrency.lockutils [req-60756992-7c62-4080-b7de-9556ca839788 req-06a5c26f-b562-42ca-b23b-b3dfd8f95701 service nova] Releasing lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.419214] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquired lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.419214] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 602.548811] env[62000]: DEBUG nova.network.neutron [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.908559] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c402876-85a4-4a4e-a2f8-da0c9f6eccb8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.917923] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815f3bcf-cb5e-4172-a1b3-74341620c28a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.954532] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa50a927-c36d-439a-9e3e-71aaf55a810b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.962088] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75762067-22bb-4474-b04c-6ef1d8d40396 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.977166] env[62000]: DEBUG nova.compute.provider_tree [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.978900] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.053812] env[62000]: INFO nova.compute.manager [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] [instance: 57cc0a22-28ff-4a24-9fea-5c6a0a699a67] Took 1.03 seconds to deallocate network for instance. [ 603.064918] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.324036] env[62000]: DEBUG nova.compute.manager [req-38141ff6-945c-4f39-896a-c2ed9177ffca req-e42a3366-8d92-4637-b38b-bd3f67f45679 service nova] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Received event network-vif-deleted-a8099270-f889-49e5-8e92-13e439ea9de1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 603.484059] env[62000]: DEBUG nova.scheduler.client.report [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.567245] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Releasing lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.567432] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.567625] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 603.567902] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f67e5fc-7150-4b40-a110-8544d2a26431 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.576698] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643b867a-ea7e-4341-873b-a3104f8569d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.598760] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6ac79769-fc49-4dd9-abae-fc2113983257 could not be found. [ 603.598984] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 603.599192] env[62000]: INFO nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Took 0.03 seconds to destroy the instance on the hypervisor. [ 603.600135] env[62000]: DEBUG oslo.service.loopingcall [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.600312] env[62000]: DEBUG nova.compute.manager [-] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.600412] env[62000]: DEBUG nova.network.neutron [-] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 603.617395] env[62000]: DEBUG nova.network.neutron [-] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 603.696867] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.697399] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.988802] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.098s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 603.989463] env[62000]: ERROR nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Traceback (most recent call last): [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.driver.spawn(context, instance, image_meta, [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] vm_ref = self.build_virtual_machine(instance, [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] vif_infos = vmwarevif.get_vif_info(self._session, [ 603.989463] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] for vif in network_info: [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self._sync_wrapper(fn, *args, **kwargs) [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.wait() [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self[:] = self._gt.wait() [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self._exit_event.wait() [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] result = hub.switch() [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 603.989807] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return self.greenlet.switch() [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] result = function(*args, **kwargs) [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] return func(*args, **kwargs) [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise e [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] nwinfo = self.network_api.allocate_for_instance( [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] created_port_ids = self._update_ports_for_instance( [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] with excutils.save_and_reraise_exception(): [ 603.990187] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] self.force_reraise() [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise self.value [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] updated_port = self._update_port( [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] _ensure_no_port_binding_failure(port) [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] raise exception.PortBindingFailed(port_id=port['id']) [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] nova.exception.PortBindingFailed: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. [ 603.990546] env[62000]: ERROR nova.compute.manager [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] [ 603.990872] env[62000]: DEBUG nova.compute.utils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 603.991512] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.132s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 603.993301] env[62000]: INFO nova.compute.claims [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 603.996731] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Build of instance c0f34d3e-39a1-4487-bd1f-ade41ef6832c was re-scheduled: Binding failed for port 16cacb86-1eef-4bc8-8bb8-19573360c52b, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 603.997177] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 603.997402] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquiring lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.997546] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Acquired lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.997704] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 604.094835] env[62000]: INFO nova.scheduler.client.report [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Deleted allocations for instance 57cc0a22-28ff-4a24-9fea-5c6a0a699a67 [ 604.119163] env[62000]: DEBUG nova.network.neutron [-] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.519089] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 604.572586] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.602503] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d696a0ee-90a2-4b12-a4f1-7704bfdda370 tempest-VolumesAssistedSnapshotsTest-356776037 tempest-VolumesAssistedSnapshotsTest-356776037-project-member] Lock "57cc0a22-28ff-4a24-9fea-5c6a0a699a67" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.856s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.621368] env[62000]: INFO nova.compute.manager [-] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Took 1.02 seconds to deallocate network for instance. [ 604.623799] env[62000]: DEBUG nova.compute.claims [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.623989] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 605.076677] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Releasing lock "refresh_cache-c0f34d3e-39a1-4487-bd1f-ade41ef6832c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.076925] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.077111] env[62000]: DEBUG nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.077412] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 605.097557] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 605.105211] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 605.508768] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a01f1fd-4b1f-4f68-afae-c5c24a67ba17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.516734] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9377292-9324-403c-a66d-6d3ddebb0bf5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.550808] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8226ce04-3c34-4e2a-b5c8-f5a8d2bdf255 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.557943] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6429491d-a48f-4493-8c76-bfeaeaa91b5d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.572281] env[62000]: DEBUG nova.compute.provider_tree [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.601813] env[62000]: DEBUG nova.network.neutron [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.624182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.076902] env[62000]: DEBUG nova.scheduler.client.report [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.105507] env[62000]: INFO nova.compute.manager [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] [instance: c0f34d3e-39a1-4487-bd1f-ade41ef6832c] Took 1.03 seconds to deallocate network for instance. [ 606.582506] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.582727] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 606.586022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.072s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.094368] env[62000]: DEBUG nova.compute.utils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.096201] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.096363] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 607.141092] env[62000]: INFO nova.scheduler.client.report [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Deleted allocations for instance c0f34d3e-39a1-4487-bd1f-ade41ef6832c [ 607.168152] env[62000]: DEBUG nova.policy [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.585492] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68bb94dd-e1a6-40a2-8813-3ad2634901e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.595301] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152e149d-7412-40b5-9fde-3e8ea9018f12 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.634666] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 607.638663] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187cbaa8-5d1d-4d46-936b-ec16e4a17ca2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.646459] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fc9809b-5ac9-4b90-b1dd-39108de9b243 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.650681] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06d58b0f-2e35-42d0-b767-91af563b1f3c tempest-ServerDiagnosticsTest-1258149472 tempest-ServerDiagnosticsTest-1258149472-project-member] Lock "c0f34d3e-39a1-4487-bd1f-ade41ef6832c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.413s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.665215] env[62000]: DEBUG nova.compute.provider_tree [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.810150] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Successfully created port: 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.153587] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.168248] env[62000]: DEBUG nova.scheduler.client.report [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.650169] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 608.675949] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.090s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.677250] env[62000]: ERROR nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Traceback (most recent call last): [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.driver.spawn(context, instance, image_meta, [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] vm_ref = self.build_virtual_machine(instance, [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.677250] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] for vif in network_info: [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self._sync_wrapper(fn, *args, **kwargs) [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.wait() [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self[:] = self._gt.wait() [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self._exit_event.wait() [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] result = hub.switch() [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.677555] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return self.greenlet.switch() [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] result = function(*args, **kwargs) [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] return func(*args, **kwargs) [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise e [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] nwinfo = self.network_api.allocate_for_instance( [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] created_port_ids = self._update_ports_for_instance( [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] with excutils.save_and_reraise_exception(): [ 608.677910] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] self.force_reraise() [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise self.value [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] updated_port = self._update_port( [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] _ensure_no_port_binding_failure(port) [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] raise exception.PortBindingFailed(port_id=port['id']) [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] nova.exception.PortBindingFailed: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. [ 608.678226] env[62000]: ERROR nova.compute.manager [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] [ 608.678831] env[62000]: DEBUG nova.compute.utils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.680731] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.349s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.681914] env[62000]: INFO nova.compute.claims [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 608.691018] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Build of instance 465d94c2-5ebf-4b05-966e-9272e2ddebb3 was re-scheduled: Binding failed for port 329e734b-bbed-49c4-ab09-1a7b00390583, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.691018] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.691315] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquiring lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.691315] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Acquired lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.691406] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 608.693311] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 608.695417] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 608.695641] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 608.695794] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 608.695971] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 608.696569] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 608.697886] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 608.698163] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 608.698335] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 608.698523] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 608.698679] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 608.698849] env[62000]: DEBUG nova.virt.hardware [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 608.699743] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac90ec30-59d1-4b7c-ab35-61c9c7c47e2b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.713944] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37c1edc-0c72-44ae-9a8c-c44e13dfba05 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.734093] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 608.920537] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.265148] env[62000]: DEBUG nova.compute.manager [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Received event network-changed-1d40dfa1-79a9-44f4-8c6b-d0982f25d04d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 609.265409] env[62000]: DEBUG nova.compute.manager [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Refreshing instance network info cache due to event network-changed-1d40dfa1-79a9-44f4-8c6b-d0982f25d04d. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 609.265640] env[62000]: DEBUG oslo_concurrency.lockutils [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] Acquiring lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.265786] env[62000]: DEBUG oslo_concurrency.lockutils [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] Acquired lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 609.265944] env[62000]: DEBUG nova.network.neutron [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Refreshing network info cache for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 609.423643] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Releasing lock "refresh_cache-465d94c2-5ebf-4b05-966e-9272e2ddebb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.423893] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 609.424084] env[62000]: DEBUG nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.424249] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 609.452385] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.587304] env[62000]: ERROR nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 609.587304] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.587304] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.587304] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.587304] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.587304] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.587304] env[62000]: ERROR nova.compute.manager raise self.value [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.587304] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 609.587304] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.587304] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 609.587721] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.587721] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 609.587721] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 609.587721] env[62000]: ERROR nova.compute.manager [ 609.587721] env[62000]: Traceback (most recent call last): [ 609.587721] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 609.587721] env[62000]: listener.cb(fileno) [ 609.587721] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.587721] env[62000]: result = function(*args, **kwargs) [ 609.587721] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.587721] env[62000]: return func(*args, **kwargs) [ 609.587721] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.587721] env[62000]: raise e [ 609.587721] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.587721] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 609.587721] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.587721] env[62000]: created_port_ids = self._update_ports_for_instance( [ 609.587721] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.587721] env[62000]: with excutils.save_and_reraise_exception(): [ 609.587721] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.587721] env[62000]: self.force_reraise() [ 609.587721] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.587721] env[62000]: raise self.value [ 609.587721] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.587721] env[62000]: updated_port = self._update_port( [ 609.587721] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.587721] env[62000]: _ensure_no_port_binding_failure(port) [ 609.587721] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.587721] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 609.588379] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 609.588379] env[62000]: Removing descriptor: 19 [ 609.588379] env[62000]: ERROR nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Traceback (most recent call last): [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] yield resources [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.driver.spawn(context, instance, image_meta, [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self._vmops.spawn(context, instance, image_meta, injected_files, [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 609.588379] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] vm_ref = self.build_virtual_machine(instance, [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] vif_infos = vmwarevif.get_vif_info(self._session, [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] for vif in network_info: [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self._sync_wrapper(fn, *args, **kwargs) [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.wait() [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self[:] = self._gt.wait() [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self._exit_event.wait() [ 609.588642] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] result = hub.switch() [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self.greenlet.switch() [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] result = function(*args, **kwargs) [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return func(*args, **kwargs) [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise e [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] nwinfo = self.network_api.allocate_for_instance( [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 609.588915] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] created_port_ids = self._update_ports_for_instance( [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] with excutils.save_and_reraise_exception(): [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.force_reraise() [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise self.value [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] updated_port = self._update_port( [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] _ensure_no_port_binding_failure(port) [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 609.589182] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise exception.PortBindingFailed(port_id=port['id']) [ 609.589424] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 609.589424] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] [ 609.589424] env[62000]: INFO nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Terminating instance [ 609.596134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 609.787431] env[62000]: DEBUG nova.network.neutron [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 609.903673] env[62000]: DEBUG nova.network.neutron [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.954878] env[62000]: DEBUG nova.network.neutron [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.185256] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0027fbc6-0178-4414-9685-b000b11b5fda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.193167] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4fc3000-adcb-44fb-9fa9-8f85ba6947bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.222596] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbff8678-569f-4fe1-b7dd-aab32917ae1d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.230120] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf6482f8-2456-448e-9595-22e741684a47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.243450] env[62000]: DEBUG nova.compute.provider_tree [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 610.406318] env[62000]: DEBUG oslo_concurrency.lockutils [req-543bd367-64bf-4b5d-904d-4af2fda1bc67 req-1355fa16-0087-4045-8933-d3f9e3de84f7 service nova] Releasing lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 610.406718] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.406896] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 610.458407] env[62000]: INFO nova.compute.manager [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] [instance: 465d94c2-5ebf-4b05-966e-9272e2ddebb3] Took 1.03 seconds to deallocate network for instance. [ 610.751092] env[62000]: DEBUG nova.scheduler.client.report [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 610.931769] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.061556] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.258687] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.578s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.259352] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 611.262466] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.858s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.262466] env[62000]: DEBUG nova.objects.instance [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lazy-loading 'resources' on Instance uuid 15409fa7-254c-435e-8080-46d3f65b2d46 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 611.290443] env[62000]: DEBUG nova.compute.manager [req-bdd71808-b1ca-4030-9d5c-f26900f5ef86 req-2fd79c43-23dd-4d7f-b5f9-64205c2dd308 service nova] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Received event network-vif-deleted-1d40dfa1-79a9-44f4-8c6b-d0982f25d04d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 611.507234] env[62000]: INFO nova.scheduler.client.report [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Deleted allocations for instance 465d94c2-5ebf-4b05-966e-9272e2ddebb3 [ 611.564746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.565656] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 611.565656] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 611.565656] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cf989ceb-ec3f-4236-afe1-9da189f42ab4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.575164] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da0148f6-ad4c-4641-96eb-0a65de530293 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.596776] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 123175a1-51a6-4463-8d90-c6b10770f905 could not be found. [ 611.597020] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 611.597847] env[62000]: INFO nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Took 0.03 seconds to destroy the instance on the hypervisor. [ 611.597847] env[62000]: DEBUG oslo.service.loopingcall [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 611.597847] env[62000]: DEBUG nova.compute.manager [-] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.597847] env[62000]: DEBUG nova.network.neutron [-] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 611.620127] env[62000]: DEBUG nova.network.neutron [-] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 611.774600] env[62000]: DEBUG nova.compute.utils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 611.774600] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 611.774600] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 611.896339] env[62000]: DEBUG nova.policy [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '147e732bf4e94249abef07fd2090ea80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45a0491d5e13450bb11b99c4fba2807f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 612.014629] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286f5c16-3ca9-4151-8d1a-851c6384a79e tempest-ServersWithSpecificFlavorTestJSON-2094742839 tempest-ServersWithSpecificFlavorTestJSON-2094742839-project-member] Lock "465d94c2-5ebf-4b05-966e-9272e2ddebb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.140s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.127188] env[62000]: DEBUG nova.network.neutron [-] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.257869] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7febdf3e-1dbb-40cc-b020-ffb9d87bc13a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.265829] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b413455-1b4b-40b3-acdb-77a28ae14ae4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.295486] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 612.298655] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9ad781-fb86-4850-8178-54730cdd1dd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.307034] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2542417-85e0-4f8d-af1e-df6fb29fb652 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.321838] env[62000]: DEBUG nova.compute.provider_tree [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 612.517226] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.629789] env[62000]: INFO nova.compute.manager [-] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Took 1.03 seconds to deallocate network for instance. [ 612.633708] env[62000]: DEBUG nova.compute.claims [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 612.633903] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.686082] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Successfully created port: 0d98f22c-730a-4d2b-bd48-346f169a9230 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 612.825848] env[62000]: DEBUG nova.scheduler.client.report [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 613.052115] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.310571] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 613.332795] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.070s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 613.335553] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.131s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 613.342299] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 613.342600] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 613.343467] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 613.343467] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 613.343467] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 613.343655] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 613.344883] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 613.345113] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 613.345432] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 613.345703] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 613.345928] env[62000]: DEBUG nova.virt.hardware [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 613.347284] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c931a0a8-fa44-4fc9-add6-8079d82bc58f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.359037] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c8d64d8-b4b8-43f6-9dc7-26a34b3b7fbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.366732] env[62000]: INFO nova.scheduler.client.report [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Deleted allocations for instance 15409fa7-254c-435e-8080-46d3f65b2d46 [ 613.706966] env[62000]: DEBUG nova.compute.manager [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Received event network-changed-0d98f22c-730a-4d2b-bd48-346f169a9230 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 613.707214] env[62000]: DEBUG nova.compute.manager [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Refreshing instance network info cache due to event network-changed-0d98f22c-730a-4d2b-bd48-346f169a9230. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 613.707404] env[62000]: DEBUG oslo_concurrency.lockutils [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] Acquiring lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.707697] env[62000]: DEBUG oslo_concurrency.lockutils [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] Acquired lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.707875] env[62000]: DEBUG nova.network.neutron [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Refreshing network info cache for port 0d98f22c-730a-4d2b-bd48-346f169a9230 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 613.848665] env[62000]: ERROR nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 613.848665] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.848665] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.848665] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.848665] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.848665] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.848665] env[62000]: ERROR nova.compute.manager raise self.value [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.848665] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.848665] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.848665] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.849025] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.849025] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.849025] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 613.849025] env[62000]: ERROR nova.compute.manager [ 613.849025] env[62000]: Traceback (most recent call last): [ 613.849025] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.849025] env[62000]: listener.cb(fileno) [ 613.849025] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.849025] env[62000]: result = function(*args, **kwargs) [ 613.849025] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.849025] env[62000]: return func(*args, **kwargs) [ 613.849025] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.849025] env[62000]: raise e [ 613.849025] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.849025] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 613.849025] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.849025] env[62000]: created_port_ids = self._update_ports_for_instance( [ 613.849025] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.849025] env[62000]: with excutils.save_and_reraise_exception(): [ 613.849025] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.849025] env[62000]: self.force_reraise() [ 613.849025] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.849025] env[62000]: raise self.value [ 613.849025] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.849025] env[62000]: updated_port = self._update_port( [ 613.849025] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.849025] env[62000]: _ensure_no_port_binding_failure(port) [ 613.849025] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.849025] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.849711] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 613.849711] env[62000]: Removing descriptor: 19 [ 613.849927] env[62000]: ERROR nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] Traceback (most recent call last): [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] yield resources [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.driver.spawn(context, instance, image_meta, [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] vm_ref = self.build_virtual_machine(instance, [ 613.849927] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] for vif in network_info: [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self._sync_wrapper(fn, *args, **kwargs) [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.wait() [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self[:] = self._gt.wait() [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self._exit_event.wait() [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.851582] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] result = hub.switch() [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self.greenlet.switch() [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] result = function(*args, **kwargs) [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return func(*args, **kwargs) [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise e [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] nwinfo = self.network_api.allocate_for_instance( [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] created_port_ids = self._update_ports_for_instance( [ 613.851999] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] with excutils.save_and_reraise_exception(): [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.force_reraise() [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise self.value [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] updated_port = self._update_port( [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] _ensure_no_port_binding_failure(port) [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise exception.PortBindingFailed(port_id=port['id']) [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 613.852331] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] [ 613.852620] env[62000]: INFO nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Terminating instance [ 613.857302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.884746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-963a12af-f547-40c7-b1de-1d66848c643e tempest-ServersAdmin275Test-1233411474 tempest-ServersAdmin275Test-1233411474-project-member] Lock "15409fa7-254c-435e-8080-46d3f65b2d46" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.367s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.235131] env[62000]: DEBUG nova.network.neutron [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 614.319567] env[62000]: DEBUG nova.network.neutron [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.342097] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a226faa-17c1-4f83-907d-4a305141bb96 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.350150] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eef02b3-7f97-4318-8a76-632980f081b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.384178] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3f9230-beda-491a-99d5-f3115db59223 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.392496] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-080ef5c0-dd3a-4ef9-ab14-5c95e7788565 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.406546] env[62000]: DEBUG nova.compute.provider_tree [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.822698] env[62000]: DEBUG oslo_concurrency.lockutils [req-475acde9-2b90-49f8-9e51-9ef0b7032f93 req-40cfc9bc-266c-429a-a158-2c9592c00cb3 service nova] Releasing lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.823126] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.823324] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 614.909865] env[62000]: DEBUG nova.scheduler.client.report [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 615.346757] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.395768] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.419225] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.084s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.419858] env[62000]: ERROR nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] Traceback (most recent call last): [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.driver.spawn(context, instance, image_meta, [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self._vmops.spawn(context, instance, image_meta, injected_files, [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] vm_ref = self.build_virtual_machine(instance, [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] vif_infos = vmwarevif.get_vif_info(self._session, [ 615.419858] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] for vif in network_info: [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self._sync_wrapper(fn, *args, **kwargs) [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.wait() [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self[:] = self._gt.wait() [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self._exit_event.wait() [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] result = hub.switch() [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 615.420319] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return self.greenlet.switch() [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] result = function(*args, **kwargs) [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] return func(*args, **kwargs) [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise e [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] nwinfo = self.network_api.allocate_for_instance( [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] created_port_ids = self._update_ports_for_instance( [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] with excutils.save_and_reraise_exception(): [ 615.420800] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] self.force_reraise() [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise self.value [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] updated_port = self._update_port( [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] _ensure_no_port_binding_failure(port) [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] raise exception.PortBindingFailed(port_id=port['id']) [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] nova.exception.PortBindingFailed: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. [ 615.421298] env[62000]: ERROR nova.compute.manager [instance: d7c25703-df04-4289-92d6-226180c43082] [ 615.421863] env[62000]: DEBUG nova.compute.utils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 615.423176] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.082s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.428023] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Build of instance d7c25703-df04-4289-92d6-226180c43082 was re-scheduled: Binding failed for port 791a6040-4ad9-4150-8630-48761277b78d, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 615.428480] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 615.428702] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.428843] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 615.428997] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 615.750145] env[62000]: DEBUG nova.compute.manager [req-3ec85224-8431-45b8-bf64-2355982c0012 req-af11e37d-eef4-454c-8416-ddba436e74fb service nova] [instance: e7739874-2088-48e3-b970-96b3d935477b] Received event network-vif-deleted-0d98f22c-730a-4d2b-bd48-346f169a9230 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 615.904043] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.908466] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.908688] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 615.909031] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f20d323-4773-4d8c-bd19-18f0dfe48990 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.922161] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0790a7-5d50-4c0f-8a49-99a3b7f3a050 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.952635] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e7739874-2088-48e3-b970-96b3d935477b could not be found. [ 615.953141] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 615.953141] env[62000]: INFO nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 615.953885] env[62000]: DEBUG oslo.service.loopingcall [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.954121] env[62000]: DEBUG nova.compute.manager [-] [instance: e7739874-2088-48e3-b970-96b3d935477b] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.954217] env[62000]: DEBUG nova.network.neutron [-] [instance: e7739874-2088-48e3-b970-96b3d935477b] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 615.983040] env[62000]: DEBUG nova.network.neutron [-] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 615.986038] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.099847] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.414926] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30642b90-8eea-40df-b74d-da8417090272 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.425772] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26577cbd-f1c4-4b02-9646-0a2dad983d86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.458041] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5e78c9-0eca-4055-ad64-12f49d4cbc5c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.465463] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12ab8ff-5745-4d4b-8613-1fe15349c32d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.478986] env[62000]: DEBUG nova.compute.provider_tree [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.489531] env[62000]: DEBUG nova.network.neutron [-] [instance: e7739874-2088-48e3-b970-96b3d935477b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.602191] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "refresh_cache-d7c25703-df04-4289-92d6-226180c43082" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.602441] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 616.602627] env[62000]: DEBUG nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.602831] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 616.623727] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 616.986942] env[62000]: DEBUG nova.scheduler.client.report [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.992332] env[62000]: INFO nova.compute.manager [-] [instance: e7739874-2088-48e3-b970-96b3d935477b] Took 1.04 seconds to deallocate network for instance. [ 616.995301] env[62000]: DEBUG nova.compute.claims [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 616.995535] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.126737] env[62000]: DEBUG nova.network.neutron [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.490087] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.490746] env[62000]: ERROR nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Traceback (most recent call last): [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.driver.spawn(context, instance, image_meta, [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] vm_ref = self.build_virtual_machine(instance, [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.490746] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] for vif in network_info: [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self._sync_wrapper(fn, *args, **kwargs) [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.wait() [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self[:] = self._gt.wait() [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self._exit_event.wait() [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] result = hub.switch() [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.491201] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return self.greenlet.switch() [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] result = function(*args, **kwargs) [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] return func(*args, **kwargs) [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise e [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] nwinfo = self.network_api.allocate_for_instance( [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] created_port_ids = self._update_ports_for_instance( [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] with excutils.save_and_reraise_exception(): [ 617.492209] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] self.force_reraise() [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise self.value [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] updated_port = self._update_port( [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] _ensure_no_port_binding_failure(port) [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] raise exception.PortBindingFailed(port_id=port['id']) [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] nova.exception.PortBindingFailed: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. [ 617.493289] env[62000]: ERROR nova.compute.manager [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] [ 617.493711] env[62000]: DEBUG nova.compute.utils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.493711] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.892s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.495622] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Build of instance a36de4b6-6928-4110-bc27-825ae58b15e5 was re-scheduled: Binding failed for port 09c8357e-8463-4f6f-8946-da3c798950eb, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.496059] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.496294] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.496438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquired lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.496598] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 617.629634] env[62000]: INFO nova.compute.manager [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: d7c25703-df04-4289-92d6-226180c43082] Took 1.03 seconds to deallocate network for instance. [ 618.031959] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.166694] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.425884] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9d168d-a9be-4803-b9f0-999a46aece3e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.433997] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-939f6e71-4372-4aea-8194-bf2e8f107a96 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.468915] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3ac6c2-8286-4f5f-8dab-41d8f5817653 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.476152] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e737c14-d940-4bf1-bd4e-5c7fb92090b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.490648] env[62000]: DEBUG nova.compute.provider_tree [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.664163] env[62000]: INFO nova.scheduler.client.report [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Deleted allocations for instance d7c25703-df04-4289-92d6-226180c43082 [ 618.671675] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Releasing lock "refresh_cache-a36de4b6-6928-4110-bc27-825ae58b15e5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.671921] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.675497] env[62000]: DEBUG nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.675602] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 618.818746] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 618.994037] env[62000]: DEBUG nova.scheduler.client.report [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.171729] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c780bc5f-8ece-485d-bfbe-cc9917acf1b5 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "d7c25703-df04-4289-92d6-226180c43082" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.671s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.322521] env[62000]: DEBUG nova.network.neutron [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.499665] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.007s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.501159] env[62000]: ERROR nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Traceback (most recent call last): [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.driver.spawn(context, instance, image_meta, [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] vm_ref = self.build_virtual_machine(instance, [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.501159] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] for vif in network_info: [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self._sync_wrapper(fn, *args, **kwargs) [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.wait() [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self[:] = self._gt.wait() [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self._exit_event.wait() [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] result = hub.switch() [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.501562] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return self.greenlet.switch() [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] result = function(*args, **kwargs) [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] return func(*args, **kwargs) [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise e [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] nwinfo = self.network_api.allocate_for_instance( [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] created_port_ids = self._update_ports_for_instance( [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] with excutils.save_and_reraise_exception(): [ 619.501917] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] self.force_reraise() [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise self.value [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] updated_port = self._update_port( [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] _ensure_no_port_binding_failure(port) [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] raise exception.PortBindingFailed(port_id=port['id']) [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] nova.exception.PortBindingFailed: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. [ 619.502224] env[62000]: ERROR nova.compute.manager [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] [ 619.502477] env[62000]: DEBUG nova.compute.utils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.508426] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.900s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.510033] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Build of instance 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f was re-scheduled: Binding failed for port c7a1e502-23ab-4f8b-81cc-95f8b94dbd79, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.510033] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.510033] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquiring lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.510033] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Acquired lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.510384] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 619.678044] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 619.829684] env[62000]: INFO nova.compute.manager [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: a36de4b6-6928-4110-bc27-825ae58b15e5] Took 1.15 seconds to deallocate network for instance. [ 620.044018] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 620.199607] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 620.231485] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.472015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76eb3e1-ce5c-4974-98c6-41c58d2621e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.479176] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333654ed-42f3-4f1d-8b59-8482e2f71305 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.515340] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc4c911c-e28c-4a73-b7a7-4e7c0c3f371c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.523399] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219f8986-68da-430e-97e6-6de9ffab5551 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.150369] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Releasing lock "refresh_cache-1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 621.150599] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 621.150775] env[62000]: DEBUG nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 621.150941] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 621.157762] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.157988] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.167988] env[62000]: DEBUG nova.compute.provider_tree [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 621.178828] env[62000]: INFO nova.scheduler.client.report [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Deleted allocations for instance a36de4b6-6928-4110-bc27-825ae58b15e5 [ 621.207630] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 621.670600] env[62000]: DEBUG nova.scheduler.client.report [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.688904] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2596b375-fede-4eab-bc7a-ad185ced88f1 tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "a36de4b6-6928-4110-bc27-825ae58b15e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.071s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.708768] env[62000]: DEBUG nova.network.neutron [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.176046] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.670s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 622.176695] env[62000]: ERROR nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Traceback (most recent call last): [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.driver.spawn(context, instance, image_meta, [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] vm_ref = self.build_virtual_machine(instance, [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] vif_infos = vmwarevif.get_vif_info(self._session, [ 622.176695] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] for vif in network_info: [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self._sync_wrapper(fn, *args, **kwargs) [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.wait() [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self[:] = self._gt.wait() [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self._exit_event.wait() [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] result = hub.switch() [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 622.177091] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return self.greenlet.switch() [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] result = function(*args, **kwargs) [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] return func(*args, **kwargs) [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise e [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] nwinfo = self.network_api.allocate_for_instance( [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] created_port_ids = self._update_ports_for_instance( [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] with excutils.save_and_reraise_exception(): [ 622.178296] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] self.force_reraise() [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise self.value [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] updated_port = self._update_port( [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] _ensure_no_port_binding_failure(port) [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] raise exception.PortBindingFailed(port_id=port['id']) [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] nova.exception.PortBindingFailed: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. [ 622.178640] env[62000]: ERROR nova.compute.manager [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] [ 622.179065] env[62000]: DEBUG nova.compute.utils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 622.180553] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.557s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.186539] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Build of instance 3bd34ec2-5253-4578-baf3-11775de18ed4 was re-scheduled: Binding failed for port ebc7822a-dc8c-477e-8899-d913cd24e66a, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 622.186639] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 622.189017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquiring lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.189017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Acquired lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.189017] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 622.190534] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.211197] env[62000]: INFO nova.compute.manager [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] [instance: 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f] Took 1.06 seconds to deallocate network for instance. [ 622.722024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.747075] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 622.866474] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.114444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf8b6da2-589c-48b1-bf62-f421f9c687ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.128036] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d6eee32-5476-4369-a3dd-316a69c0f10e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.162587] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d40312b-807a-4461-9dde-0078c30fa138 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.171502] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35acf268-63a0-417a-9ba6-97165cd84442 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.188591] env[62000]: DEBUG nova.compute.provider_tree [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.256331] env[62000]: INFO nova.scheduler.client.report [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Deleted allocations for instance 1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f [ 623.371690] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Releasing lock "refresh_cache-3bd34ec2-5253-4578-baf3-11775de18ed4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.371964] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 623.372170] env[62000]: DEBUG nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 623.372337] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 623.394106] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 623.695661] env[62000]: DEBUG nova.scheduler.client.report [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.768038] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b88a9f4-b415-474a-8e87-9a4f22a7d89c tempest-ImagesOneServerNegativeTestJSON-184562005 tempest-ImagesOneServerNegativeTestJSON-184562005-project-member] Lock "1575bafd-50f9-43d8-bec9-dcc9b6e0fc9f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.569s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.815132] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "4e617bb5-84e4-4505-99e3-61289826f511" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.815375] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.903204] env[62000]: DEBUG nova.network.neutron [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.201563] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.201563] env[62000]: ERROR nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Traceback (most recent call last): [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.driver.spawn(context, instance, image_meta, [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self._vmops.spawn(context, instance, image_meta, injected_files, [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 624.201563] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] vm_ref = self.build_virtual_machine(instance, [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] vif_infos = vmwarevif.get_vif_info(self._session, [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] for vif in network_info: [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self._sync_wrapper(fn, *args, **kwargs) [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.wait() [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self[:] = self._gt.wait() [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self._exit_event.wait() [ 624.201992] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] result = hub.switch() [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return self.greenlet.switch() [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] result = function(*args, **kwargs) [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] return func(*args, **kwargs) [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise e [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] nwinfo = self.network_api.allocate_for_instance( [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 624.202371] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] created_port_ids = self._update_ports_for_instance( [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] with excutils.save_and_reraise_exception(): [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] self.force_reraise() [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise self.value [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] updated_port = self._update_port( [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] _ensure_no_port_binding_failure(port) [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 624.202684] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] raise exception.PortBindingFailed(port_id=port['id']) [ 624.203058] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] nova.exception.PortBindingFailed: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. [ 624.203058] env[62000]: ERROR nova.compute.manager [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] [ 624.203058] env[62000]: DEBUG nova.compute.utils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 624.207711] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Build of instance 6ac79769-fc49-4dd9-abae-fc2113983257 was re-scheduled: Binding failed for port a8099270-f889-49e5-8e92-13e439ea9de1, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 624.208205] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 624.208712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquiring lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 624.208712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Acquired lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 624.208865] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 624.209942] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.586s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.213382] env[62000]: INFO nova.compute.claims [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.275403] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.404269] env[62000]: INFO nova.compute.manager [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] [instance: 3bd34ec2-5253-4578-baf3-11775de18ed4] Took 1.03 seconds to deallocate network for instance. [ 624.469502] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "af49714d-8e50-4159-96a5-cf8f70580471" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.469502] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.744663] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 624.801591] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.840336] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.344637] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Releasing lock "refresh_cache-6ac79769-fc49-4dd9-abae-fc2113983257" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 625.344884] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 625.345084] env[62000]: DEBUG nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 625.345253] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 625.371307] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 625.446570] env[62000]: INFO nova.scheduler.client.report [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Deleted allocations for instance 3bd34ec2-5253-4578-baf3-11775de18ed4 [ 625.702370] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f02007c1-5555-4564-ad91-98e293ab3cdc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.711224] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d8bcb6-dd1c-4948-b9e9-885f7ec7d9e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.742467] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bf8d3b-bc46-4126-82ab-e4190ea9b366 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.752072] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5300a1-8349-453b-b370-08a23e06c33d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.765484] env[62000]: DEBUG nova.compute.provider_tree [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.877898] env[62000]: DEBUG nova.network.neutron [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.954453] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddfdf8d8-f572-40a6-81ad-280b4f900e6e tempest-ServerRescueTestJSON-828137247 tempest-ServerRescueTestJSON-828137247-project-member] Lock "3bd34ec2-5253-4578-baf3-11775de18ed4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 69.463s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.268870] env[62000]: DEBUG nova.scheduler.client.report [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.380997] env[62000]: INFO nova.compute.manager [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] [instance: 6ac79769-fc49-4dd9-abae-fc2113983257] Took 1.04 seconds to deallocate network for instance. [ 626.457047] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 626.774723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.564s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.775352] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.778148] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.085s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.780396] env[62000]: INFO nova.compute.claims [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 626.981617] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.286246] env[62000]: DEBUG nova.compute.utils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.288770] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.288941] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 627.336881] env[62000]: DEBUG nova.policy [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dd622b8d6ca74338acb1932d966b2b1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9bae70aa236a4629b801902c036cadb5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.410380] env[62000]: INFO nova.scheduler.client.report [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Deleted allocations for instance 6ac79769-fc49-4dd9-abae-fc2113983257 [ 627.626295] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Successfully created port: 19a392f0-dc37-4c74-b21e-210b182b9d05 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.789988] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.918440] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8435388-e724-4dcf-91fe-4db5191579b1 tempest-ServersNegativeTestJSON-54863873 tempest-ServersNegativeTestJSON-54863873-project-member] Lock "6ac79769-fc49-4dd9-abae-fc2113983257" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 70.891s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.228392] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dfc2aef-9e97-4dae-a979-9f460775d076 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.236309] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407d62fe-b80b-4480-a594-ef6f330e3128 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.275070] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31f31248-7a9b-4251-af1a-ca5bc3b4d60b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.280373] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40ab525-0981-4d2e-93ec-24b0e69c5cdb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.296267] env[62000]: DEBUG nova.compute.provider_tree [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.425524] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 628.806064] env[62000]: DEBUG nova.scheduler.client.report [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.810343] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.842789] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.843282] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.843502] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.843704] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.843849] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.844860] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.844860] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.844860] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.844860] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.844860] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.845063] env[62000]: DEBUG nova.virt.hardware [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.845950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28553fc-78c4-4479-93d3-3337d26c5273 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.856909] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2f77cc-bbbe-4292-8941-09dd957eee88 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.952738] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.316189] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.538s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.316712] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 629.320069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.686s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.436215] env[62000]: DEBUG nova.compute.manager [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Received event network-changed-19a392f0-dc37-4c74-b21e-210b182b9d05 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 629.436429] env[62000]: DEBUG nova.compute.manager [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Refreshing instance network info cache due to event network-changed-19a392f0-dc37-4c74-b21e-210b182b9d05. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 629.436638] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] Acquiring lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.436778] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] Acquired lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.436939] env[62000]: DEBUG nova.network.neutron [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Refreshing network info cache for port 19a392f0-dc37-4c74-b21e-210b182b9d05 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 629.826370] env[62000]: DEBUG nova.compute.utils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 629.827797] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 629.830127] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 629.884894] env[62000]: ERROR nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 629.884894] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.884894] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.884894] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.884894] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.884894] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.884894] env[62000]: ERROR nova.compute.manager raise self.value [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.884894] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 629.884894] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.884894] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 629.885405] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.885405] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 629.885405] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 629.885405] env[62000]: ERROR nova.compute.manager [ 629.885405] env[62000]: Traceback (most recent call last): [ 629.885405] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 629.885405] env[62000]: listener.cb(fileno) [ 629.885405] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.885405] env[62000]: result = function(*args, **kwargs) [ 629.885405] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.885405] env[62000]: return func(*args, **kwargs) [ 629.885405] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.885405] env[62000]: raise e [ 629.885405] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.885405] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 629.885405] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.885405] env[62000]: created_port_ids = self._update_ports_for_instance( [ 629.885405] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.885405] env[62000]: with excutils.save_and_reraise_exception(): [ 629.885405] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.885405] env[62000]: self.force_reraise() [ 629.885405] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.885405] env[62000]: raise self.value [ 629.885405] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.885405] env[62000]: updated_port = self._update_port( [ 629.885405] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.885405] env[62000]: _ensure_no_port_binding_failure(port) [ 629.885405] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.885405] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 629.886197] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 629.886197] env[62000]: Removing descriptor: 19 [ 629.886197] env[62000]: ERROR nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Traceback (most recent call last): [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] yield resources [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.driver.spawn(context, instance, image_meta, [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.886197] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] vm_ref = self.build_virtual_machine(instance, [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] for vif in network_info: [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self._sync_wrapper(fn, *args, **kwargs) [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.wait() [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self[:] = self._gt.wait() [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self._exit_event.wait() [ 629.886542] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] result = hub.switch() [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self.greenlet.switch() [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] result = function(*args, **kwargs) [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return func(*args, **kwargs) [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise e [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] nwinfo = self.network_api.allocate_for_instance( [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 629.886884] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] created_port_ids = self._update_ports_for_instance( [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] with excutils.save_and_reraise_exception(): [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.force_reraise() [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise self.value [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] updated_port = self._update_port( [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] _ensure_no_port_binding_failure(port) [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.887194] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise exception.PortBindingFailed(port_id=port['id']) [ 629.887825] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 629.887825] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] [ 629.887825] env[62000]: INFO nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Terminating instance [ 629.888307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquiring lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.961512] env[62000]: DEBUG nova.policy [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '187d789b014a4e6ba7fdb719f1ad0db0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9f40435bd544efcada1427af9fc8596', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 629.991845] env[62000]: DEBUG nova.network.neutron [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 630.239018] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daadc85a-5277-401d-8750-57f21497f8c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.245074] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f07f57e-3a97-44a2-ade8-bfd5b858f400 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.277543] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e37351c-6773-4757-8b90-13046f218de5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.285326] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66a94fb0-c0d4-4f3a-95e6-1095aa4bbc56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.301706] env[62000]: DEBUG nova.compute.provider_tree [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 630.336439] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 630.398809] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.398809] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.417475] env[62000]: DEBUG nova.network.neutron [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.435514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.436161] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.786017] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Successfully created port: 5a465687-beb6-4d8f-94cb-00f4abada5e1 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 630.807360] env[62000]: DEBUG nova.scheduler.client.report [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.903275] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 630.903441] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 630.903563] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Rebuilding the list of instances to heal {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 630.919742] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef213b1b-ff99-4a32-80c6-b94354c2bfbe req-34761e90-037c-4a6e-8fe1-e1f39f97a5b6 service nova] Releasing lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 630.920388] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquired lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.920573] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.310771] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.991s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.311435] env[62000]: ERROR nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Traceback (most recent call last): [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.driver.spawn(context, instance, image_meta, [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] vm_ref = self.build_virtual_machine(instance, [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.311435] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] for vif in network_info: [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self._sync_wrapper(fn, *args, **kwargs) [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.wait() [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self[:] = self._gt.wait() [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self._exit_event.wait() [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] result = hub.switch() [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.311801] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return self.greenlet.switch() [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] result = function(*args, **kwargs) [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] return func(*args, **kwargs) [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise e [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] nwinfo = self.network_api.allocate_for_instance( [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] created_port_ids = self._update_ports_for_instance( [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] with excutils.save_and_reraise_exception(): [ 631.312135] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] self.force_reraise() [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise self.value [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] updated_port = self._update_port( [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] _ensure_no_port_binding_failure(port) [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] raise exception.PortBindingFailed(port_id=port['id']) [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] nova.exception.PortBindingFailed: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. [ 631.312470] env[62000]: ERROR nova.compute.manager [instance: 123175a1-51a6-4463-8d90-c6b10770f905] [ 631.312756] env[62000]: DEBUG nova.compute.utils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.313475] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.261s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.315029] env[62000]: INFO nova.compute.claims [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.321584] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Build of instance 123175a1-51a6-4463-8d90-c6b10770f905 was re-scheduled: Binding failed for port 1d40dfa1-79a9-44f4-8c6b-d0982f25d04d, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.321584] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.321584] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.321584] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.321734] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 631.345197] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 631.381823] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 631.382236] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 631.383048] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 631.383048] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 631.383048] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 631.383048] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 631.383261] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 631.383486] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 631.383702] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 631.383878] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 631.384069] env[62000]: DEBUG nova.virt.hardware [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 631.385271] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4333aa1-d814-4ecc-ae5b-a64903548e67 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.396157] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51ccec8f-c6a5-4eb2-9a9f-89d07f021596 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 631.413208] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: e7739874-2088-48e3-b970-96b3d935477b] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 631.413378] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 631.413509] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 631.413634] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Didn't find any instances for network info cache update. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 631.414016] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.414213] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.414721] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.414918] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.415109] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.415231] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.415359] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 631.415531] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 631.471450] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.505935] env[62000]: DEBUG nova.compute.manager [req-e5267040-a8b6-433d-9054-ddeddf6bac37 req-8897e566-a07e-4235-8604-a8f9d6c32c9d service nova] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Received event network-vif-deleted-19a392f0-dc37-4c74-b21e-210b182b9d05 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 631.592561] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.860391] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 631.918977] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 632.098164] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Releasing lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.098164] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 632.098164] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 632.098164] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-21dc9cb1-1a29-4c46-a5ae-6948292cf8b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.101054] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.110987] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e52affa3-daa4-4e2a-9e21-631a1a98f853 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.139721] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6186f74-8a8e-4145-8126-7048ad29673f could not be found. [ 632.140027] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 632.140220] env[62000]: INFO nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 632.140460] env[62000]: DEBUG oslo.service.loopingcall [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 632.140695] env[62000]: DEBUG nova.compute.manager [-] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.140800] env[62000]: DEBUG nova.network.neutron [-] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.165873] env[62000]: DEBUG nova.network.neutron [-] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.376439] env[62000]: ERROR nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 632.376439] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.376439] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.376439] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.376439] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.376439] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.376439] env[62000]: ERROR nova.compute.manager raise self.value [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.376439] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 632.376439] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.376439] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 632.377026] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.377026] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 632.377026] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 632.377026] env[62000]: ERROR nova.compute.manager [ 632.377026] env[62000]: Traceback (most recent call last): [ 632.377026] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 632.377026] env[62000]: listener.cb(fileno) [ 632.377026] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.377026] env[62000]: result = function(*args, **kwargs) [ 632.377026] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.377026] env[62000]: return func(*args, **kwargs) [ 632.377026] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.377026] env[62000]: raise e [ 632.377026] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.377026] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 632.377026] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.377026] env[62000]: created_port_ids = self._update_ports_for_instance( [ 632.377026] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.377026] env[62000]: with excutils.save_and_reraise_exception(): [ 632.377026] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.377026] env[62000]: self.force_reraise() [ 632.377026] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.377026] env[62000]: raise self.value [ 632.377026] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.377026] env[62000]: updated_port = self._update_port( [ 632.377026] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.377026] env[62000]: _ensure_no_port_binding_failure(port) [ 632.377026] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.377026] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 632.377699] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 632.377699] env[62000]: Removing descriptor: 14 [ 632.377699] env[62000]: ERROR nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Traceback (most recent call last): [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] yield resources [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.driver.spawn(context, instance, image_meta, [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 632.377699] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] vm_ref = self.build_virtual_machine(instance, [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] for vif in network_info: [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self._sync_wrapper(fn, *args, **kwargs) [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.wait() [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self[:] = self._gt.wait() [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self._exit_event.wait() [ 632.377969] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] result = hub.switch() [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self.greenlet.switch() [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] result = function(*args, **kwargs) [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return func(*args, **kwargs) [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise e [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] nwinfo = self.network_api.allocate_for_instance( [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 632.378267] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] created_port_ids = self._update_ports_for_instance( [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] with excutils.save_and_reraise_exception(): [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.force_reraise() [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise self.value [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] updated_port = self._update_port( [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] _ensure_no_port_binding_failure(port) [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 632.378622] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise exception.PortBindingFailed(port_id=port['id']) [ 632.378899] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 632.378899] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] [ 632.378899] env[62000]: INFO nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Terminating instance [ 632.380122] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 632.380358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquired lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.380593] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 632.605453] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-123175a1-51a6-4463-8d90-c6b10770f905" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.608683] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 632.608683] env[62000]: DEBUG nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 632.608683] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 632.630349] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.668346] env[62000]: DEBUG nova.network.neutron [-] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.733317] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b98ce08f-327b-4116-834c-08dd4db798e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.740852] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91a6a07b-1be5-4260-8024-f4c999f793ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.773756] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96512305-93c5-4b43-8757-9ada930870f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.782346] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0518905-bac8-4c39-bf44-3a211335e7a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.796762] env[62000]: DEBUG nova.compute.provider_tree [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.910997] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 632.997210] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.133474] env[62000]: DEBUG nova.network.neutron [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.171675] env[62000]: INFO nova.compute.manager [-] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Took 1.03 seconds to deallocate network for instance. [ 633.175842] env[62000]: DEBUG nova.compute.claims [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 633.176034] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.300471] env[62000]: DEBUG nova.scheduler.client.report [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 633.386301] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.386584] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.499166] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Releasing lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.499622] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 633.499818] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 633.500130] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-56708da3-8ec2-42b0-b8c1-d7892db0a923 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.509093] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fecc4782-2097-49f9-8659-c1a527ae2391 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.533269] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7396796d-77e5-4364-b889-9cf3cbdad3c9 could not be found. [ 633.533883] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 633.534118] env[62000]: INFO nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 633.534469] env[62000]: DEBUG oslo.service.loopingcall [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 633.535702] env[62000]: DEBUG nova.compute.manager [-] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.535875] env[62000]: DEBUG nova.network.neutron [-] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 633.538144] env[62000]: DEBUG nova.compute.manager [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Received event network-changed-5a465687-beb6-4d8f-94cb-00f4abada5e1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 633.538325] env[62000]: DEBUG nova.compute.manager [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Refreshing instance network info cache due to event network-changed-5a465687-beb6-4d8f-94cb-00f4abada5e1. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 633.538559] env[62000]: DEBUG oslo_concurrency.lockutils [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] Acquiring lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.538697] env[62000]: DEBUG oslo_concurrency.lockutils [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] Acquired lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.538804] env[62000]: DEBUG nova.network.neutron [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Refreshing network info cache for port 5a465687-beb6-4d8f-94cb-00f4abada5e1 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 633.554303] env[62000]: DEBUG nova.network.neutron [-] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 633.636060] env[62000]: INFO nova.compute.manager [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 123175a1-51a6-4463-8d90-c6b10770f905] Took 1.03 seconds to deallocate network for instance. [ 633.807798] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.808330] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.810942] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.815s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.056744] env[62000]: DEBUG nova.network.neutron [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 634.058409] env[62000]: DEBUG nova.network.neutron [-] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.149398] env[62000]: DEBUG nova.network.neutron [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.315700] env[62000]: DEBUG nova.compute.utils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.320137] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.320426] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 634.370184] env[62000]: DEBUG nova.policy [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6a4afad6b3943eca890b41032f20c3d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'af0cfb674a7e46aaa502b61838d8bc97', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.561085] env[62000]: INFO nova.compute.manager [-] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Took 1.03 seconds to deallocate network for instance. [ 634.563814] env[62000]: DEBUG nova.compute.claims [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 634.563909] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.651601] env[62000]: DEBUG oslo_concurrency.lockutils [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] Releasing lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.651862] env[62000]: DEBUG nova.compute.manager [req-e88a0459-c16b-44a1-9685-bb20c6fa95b2 req-5776be7c-dff9-41c9-84bf-01754e502930 service nova] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Received event network-vif-deleted-5a465687-beb6-4d8f-94cb-00f4abada5e1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 634.669392] env[62000]: INFO nova.scheduler.client.report [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted allocations for instance 123175a1-51a6-4463-8d90-c6b10770f905 [ 634.674874] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aca1a07-1eab-416e-80ce-867ab0534d79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.686176] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b31d99-4b0b-4781-9920-849f2d238747 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.721281] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Successfully created port: b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.725669] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b410270-eea6-4457-8d2e-6ca0789fafa8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.734250] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cda37f1-beb3-4fad-92fb-f7cc87a3fc47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.747991] env[62000]: DEBUG nova.compute.provider_tree [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 634.820922] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.181514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e40e692d-e9a8-4cb8-9a1d-4a73626fb34d tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "123175a1-51a6-4463-8d90-c6b10770f905" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.950s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.251125] env[62000]: DEBUG nova.scheduler.client.report [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.578535] env[62000]: DEBUG nova.compute.manager [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Received event network-changed-b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 635.579077] env[62000]: DEBUG nova.compute.manager [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Refreshing instance network info cache due to event network-changed-b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 635.579077] env[62000]: DEBUG oslo_concurrency.lockutils [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] Acquiring lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.579077] env[62000]: DEBUG oslo_concurrency.lockutils [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] Acquired lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.579261] env[62000]: DEBUG nova.network.neutron [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Refreshing network info cache for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 635.684546] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 635.737657] env[62000]: ERROR nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 635.737657] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.737657] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.737657] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.737657] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.737657] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.737657] env[62000]: ERROR nova.compute.manager raise self.value [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.737657] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.737657] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.737657] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.738081] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.738081] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.738081] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 635.738081] env[62000]: ERROR nova.compute.manager [ 635.738081] env[62000]: Traceback (most recent call last): [ 635.738081] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.738081] env[62000]: listener.cb(fileno) [ 635.738081] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.738081] env[62000]: result = function(*args, **kwargs) [ 635.738081] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.738081] env[62000]: return func(*args, **kwargs) [ 635.738081] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.738081] env[62000]: raise e [ 635.738081] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.738081] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 635.738081] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.738081] env[62000]: created_port_ids = self._update_ports_for_instance( [ 635.738081] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.738081] env[62000]: with excutils.save_and_reraise_exception(): [ 635.738081] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.738081] env[62000]: self.force_reraise() [ 635.738081] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.738081] env[62000]: raise self.value [ 635.738081] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.738081] env[62000]: updated_port = self._update_port( [ 635.738081] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.738081] env[62000]: _ensure_no_port_binding_failure(port) [ 635.738081] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.738081] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.738854] env[62000]: nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 635.738854] env[62000]: Removing descriptor: 14 [ 635.757257] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.946s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.757373] env[62000]: ERROR nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] Traceback (most recent call last): [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.driver.spawn(context, instance, image_meta, [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] vm_ref = self.build_virtual_machine(instance, [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.757373] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] for vif in network_info: [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self._sync_wrapper(fn, *args, **kwargs) [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.wait() [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self[:] = self._gt.wait() [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self._exit_event.wait() [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] result = hub.switch() [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.757630] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return self.greenlet.switch() [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] result = function(*args, **kwargs) [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] return func(*args, **kwargs) [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise e [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] nwinfo = self.network_api.allocate_for_instance( [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] created_port_ids = self._update_ports_for_instance( [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] with excutils.save_and_reraise_exception(): [ 635.757928] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] self.force_reraise() [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise self.value [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] updated_port = self._update_port( [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] _ensure_no_port_binding_failure(port) [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] raise exception.PortBindingFailed(port_id=port['id']) [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] nova.exception.PortBindingFailed: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. [ 635.758233] env[62000]: ERROR nova.compute.manager [instance: e7739874-2088-48e3-b970-96b3d935477b] [ 635.758485] env[62000]: DEBUG nova.compute.utils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 635.759305] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.560s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.761449] env[62000]: INFO nova.compute.claims [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 635.764058] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Build of instance e7739874-2088-48e3-b970-96b3d935477b was re-scheduled: Binding failed for port 0d98f22c-730a-4d2b-bd48-346f169a9230, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 635.764557] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 635.764788] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.764935] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.766812] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 635.831615] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.855734] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.856034] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.856237] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.856424] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.856571] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.856717] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.856924] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.857103] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.857274] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.857437] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.857607] env[62000]: DEBUG nova.virt.hardware [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.858490] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fab5006-a81b-4750-855f-e3a0e93b39d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.867506] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77367c8d-1c01-413f-8763-56ff210c0fcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.882425] env[62000]: ERROR nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Traceback (most recent call last): [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] yield resources [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.driver.spawn(context, instance, image_meta, [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] vm_ref = self.build_virtual_machine(instance, [ 635.882425] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] for vif in network_info: [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return self._sync_wrapper(fn, *args, **kwargs) [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.wait() [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self[:] = self._gt.wait() [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return self._exit_event.wait() [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 635.882714] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] current.throw(*self._exc) [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] result = function(*args, **kwargs) [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return func(*args, **kwargs) [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise e [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] nwinfo = self.network_api.allocate_for_instance( [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] created_port_ids = self._update_ports_for_instance( [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] with excutils.save_and_reraise_exception(): [ 635.883122] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.force_reraise() [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise self.value [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] updated_port = self._update_port( [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] _ensure_no_port_binding_failure(port) [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise exception.PortBindingFailed(port_id=port['id']) [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 635.883451] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] [ 635.883451] env[62000]: INFO nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Terminating instance [ 635.884845] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquiring lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.103261] env[62000]: DEBUG nova.network.neutron [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.207622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.369170] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 636.401635] env[62000]: DEBUG nova.network.neutron [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.439909] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.904813] env[62000]: DEBUG oslo_concurrency.lockutils [req-8444faca-1405-4570-b5df-fc7665ce7773 req-ade9350d-8caf-43a4-bb99-9607f53936a8 service nova] Releasing lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.905263] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquired lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.905446] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 636.942625] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-e7739874-2088-48e3-b970-96b3d935477b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.942870] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 636.943140] env[62000]: DEBUG nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.943329] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 636.970095] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.138468] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e174a133-719a-4561-aae3-b73607766d7d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.147290] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521dc1e5-4e5c-43ac-80b3-7f543f1df162 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.178682] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f861eda-660f-4c43-a7b7-1202be9771ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.186572] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad05bba0-4b98-4918-b858-8b8c710d1d1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.203460] env[62000]: DEBUG nova.compute.provider_tree [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.425267] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 637.474800] env[62000]: DEBUG nova.network.neutron [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.523236] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.604721] env[62000]: DEBUG nova.compute.manager [req-dcb45891-3c6d-4e38-8731-b98cb68fbb56 req-183f9496-9110-4b4d-a79f-20a598bb5e82 service nova] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Received event network-vif-deleted-b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 637.706242] env[62000]: DEBUG nova.scheduler.client.report [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.882733] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.882997] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.976582] env[62000]: INFO nova.compute.manager [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: e7739874-2088-48e3-b970-96b3d935477b] Took 1.03 seconds to deallocate network for instance. [ 638.025805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Releasing lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.026248] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 638.026443] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 638.026781] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5f45e371-a58f-465d-9672-5b12759b1b14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.035766] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1d2826-9f4c-44c3-9a62-716a9c88c3dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 638.057637] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 289e3403-c90d-40f0-9e18-1d10c0f8aec4 could not be found. [ 638.057850] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 638.058042] env[62000]: INFO nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 638.058287] env[62000]: DEBUG oslo.service.loopingcall [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 638.058497] env[62000]: DEBUG nova.compute.manager [-] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.058593] env[62000]: DEBUG nova.network.neutron [-] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 638.073588] env[62000]: DEBUG nova.network.neutron [-] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 638.211277] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.211824] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 638.214735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.495s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.216123] env[62000]: INFO nova.compute.claims [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 638.575506] env[62000]: DEBUG nova.network.neutron [-] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.723128] env[62000]: DEBUG nova.compute.utils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 638.724642] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 638.724835] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 638.775592] env[62000]: DEBUG nova.policy [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9ab8430f2275456a9b16bb2c3573e6f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a070b994c52040839c485058d9500eb9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 639.014329] env[62000]: INFO nova.scheduler.client.report [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Deleted allocations for instance e7739874-2088-48e3-b970-96b3d935477b [ 639.068601] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Successfully created port: ee1823ff-c772-4e4f-ba45-7b5ec4dde1df {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 639.078448] env[62000]: INFO nova.compute.manager [-] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Took 1.02 seconds to deallocate network for instance. [ 639.080852] env[62000]: DEBUG nova.compute.claims [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 639.081741] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.230341] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 639.522075] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e9946-20eb-43c8-8004-06a9902f0c34 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "e7739874-2088-48e3-b970-96b3d935477b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 82.164s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 639.642038] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b5c9dd8-6a80-481d-9411-2009acf5c3a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.650356] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edb5604-2d5d-4455-ba53-4c3025ee8502 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.688842] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0598ebfa-7609-475c-ad50-c7843327b402 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.698120] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8a902a-7136-4a16-9488-92ec5d2e683b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.712052] env[62000]: DEBUG nova.compute.provider_tree [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.849307] env[62000]: DEBUG nova.compute.manager [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Received event network-changed-ee1823ff-c772-4e4f-ba45-7b5ec4dde1df {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 639.853020] env[62000]: DEBUG nova.compute.manager [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Refreshing instance network info cache due to event network-changed-ee1823ff-c772-4e4f-ba45-7b5ec4dde1df. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 639.853020] env[62000]: DEBUG oslo_concurrency.lockutils [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] Acquiring lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.853020] env[62000]: DEBUG oslo_concurrency.lockutils [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] Acquired lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.853020] env[62000]: DEBUG nova.network.neutron [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Refreshing network info cache for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 640.027500] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 640.067860] env[62000]: ERROR nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 640.067860] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.067860] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.067860] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.067860] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.067860] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.067860] env[62000]: ERROR nova.compute.manager raise self.value [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.067860] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 640.067860] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.067860] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 640.068261] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.068261] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 640.068261] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 640.068261] env[62000]: ERROR nova.compute.manager [ 640.068261] env[62000]: Traceback (most recent call last): [ 640.068261] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 640.068261] env[62000]: listener.cb(fileno) [ 640.068261] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.068261] env[62000]: result = function(*args, **kwargs) [ 640.068261] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.068261] env[62000]: return func(*args, **kwargs) [ 640.068261] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.068261] env[62000]: raise e [ 640.068261] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.068261] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 640.068261] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.068261] env[62000]: created_port_ids = self._update_ports_for_instance( [ 640.068261] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.068261] env[62000]: with excutils.save_and_reraise_exception(): [ 640.068261] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.068261] env[62000]: self.force_reraise() [ 640.068261] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.068261] env[62000]: raise self.value [ 640.068261] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.068261] env[62000]: updated_port = self._update_port( [ 640.068261] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.068261] env[62000]: _ensure_no_port_binding_failure(port) [ 640.068261] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.068261] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 640.069056] env[62000]: nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 640.069056] env[62000]: Removing descriptor: 19 [ 640.215349] env[62000]: DEBUG nova.scheduler.client.report [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.245859] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 640.273570] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 640.273570] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 640.273872] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 640.273903] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 640.274085] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 640.274191] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 640.274408] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 640.274570] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 640.274748] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 640.274914] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 640.275102] env[62000]: DEBUG nova.virt.hardware [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 640.275977] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6175dc95-a2a6-4340-b224-34f9b96b6d41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.284116] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad1172f-db6c-4a1d-9457-28ac04a5bc80 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.298565] env[62000]: ERROR nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Traceback (most recent call last): [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] yield resources [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.driver.spawn(context, instance, image_meta, [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] vm_ref = self.build_virtual_machine(instance, [ 640.298565] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] for vif in network_info: [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return self._sync_wrapper(fn, *args, **kwargs) [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.wait() [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self[:] = self._gt.wait() [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return self._exit_event.wait() [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 640.298978] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] current.throw(*self._exc) [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] result = function(*args, **kwargs) [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return func(*args, **kwargs) [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise e [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] nwinfo = self.network_api.allocate_for_instance( [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] created_port_ids = self._update_ports_for_instance( [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] with excutils.save_and_reraise_exception(): [ 640.299311] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.force_reraise() [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise self.value [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] updated_port = self._update_port( [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] _ensure_no_port_binding_failure(port) [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise exception.PortBindingFailed(port_id=port['id']) [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 640.299645] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] [ 640.299645] env[62000]: INFO nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Terminating instance [ 640.301522] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.366893] env[62000]: DEBUG nova.network.neutron [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 640.422476] env[62000]: DEBUG nova.network.neutron [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.544521] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.722154] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.507s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.722936] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.725577] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.924s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.727069] env[62000]: INFO nova.compute.claims [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.928358] env[62000]: DEBUG oslo_concurrency.lockutils [req-11a67920-d89e-4328-9464-6a3c7bcfae63 req-33862316-add1-4995-8bd7-16a5e7ff211e service nova] Releasing lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.928834] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquired lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.929058] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 640.958194] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "f3a6527c-777e-4a07-9482-598de15d4eb3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.958428] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.231023] env[62000]: DEBUG nova.compute.utils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 641.234933] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 641.235174] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 641.285107] env[62000]: DEBUG nova.policy [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0014136775f42db92d5dbc4f09906ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9eb068823bc74cc99499568ebd94650d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 641.450923] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 641.559843] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.634012] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Successfully created port: 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.741512] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.884285] env[62000]: DEBUG nova.compute.manager [req-5e0251ba-f75c-4f90-b725-b0ddd7e7a0d1 req-29c33b75-2fe2-4b21-889b-03640bbcdd5d service nova] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Received event network-vif-deleted-ee1823ff-c772-4e4f-ba45-7b5ec4dde1df {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 642.063013] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Releasing lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 642.063536] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 642.063943] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 642.064932] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4bcbecca-315d-42d1-831f-84c4147cbdb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.077589] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a0f4638-f40f-4c46-a44f-7e263fd81551 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.101395] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ab167fa4-062b-4137-801c-9b57a17b20e2 could not be found. [ 642.101395] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 642.101395] env[62000]: INFO nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 642.101395] env[62000]: DEBUG oslo.service.loopingcall [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 642.103213] env[62000]: DEBUG nova.compute.manager [-] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 642.103288] env[62000]: DEBUG nova.network.neutron [-] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 642.126603] env[62000]: DEBUG nova.network.neutron [-] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 642.184385] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e989df-8053-47e1-86c1-6b5713b56404 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.191750] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ba4e2d-d3b3-431f-94b8-c4ba26325497 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.222511] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0a408c-b749-473b-9c52-e033d93ba1e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.229921] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e99877aa-5445-439e-ad41-5efef8b78a5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.243415] env[62000]: DEBUG nova.compute.provider_tree [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 642.578476] env[62000]: ERROR nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 642.578476] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.578476] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.578476] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.578476] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.578476] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.578476] env[62000]: ERROR nova.compute.manager raise self.value [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.578476] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.578476] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.578476] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.578989] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.578989] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.578989] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 642.578989] env[62000]: ERROR nova.compute.manager [ 642.578989] env[62000]: Traceback (most recent call last): [ 642.578989] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.578989] env[62000]: listener.cb(fileno) [ 642.578989] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.578989] env[62000]: result = function(*args, **kwargs) [ 642.578989] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.578989] env[62000]: return func(*args, **kwargs) [ 642.578989] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.578989] env[62000]: raise e [ 642.578989] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.578989] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 642.578989] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.578989] env[62000]: created_port_ids = self._update_ports_for_instance( [ 642.578989] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.578989] env[62000]: with excutils.save_and_reraise_exception(): [ 642.578989] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.578989] env[62000]: self.force_reraise() [ 642.578989] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.578989] env[62000]: raise self.value [ 642.578989] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.578989] env[62000]: updated_port = self._update_port( [ 642.578989] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.578989] env[62000]: _ensure_no_port_binding_failure(port) [ 642.578989] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.578989] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.579981] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 642.579981] env[62000]: Removing descriptor: 19 [ 642.632366] env[62000]: DEBUG nova.network.neutron [-] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.746109] env[62000]: DEBUG nova.scheduler.client.report [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.751713] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.778046] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.778046] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.778046] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.778313] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.778313] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.778313] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.778313] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.778313] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.778467] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.778467] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.778642] env[62000]: DEBUG nova.virt.hardware [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.779859] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a2da60f-5342-42ea-9188-693d5d027bac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.789102] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9e47a1-aaab-43b2-aaf4-157b7b97469f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.803483] env[62000]: ERROR nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Traceback (most recent call last): [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] yield resources [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.driver.spawn(context, instance, image_meta, [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] vm_ref = self.build_virtual_machine(instance, [ 642.803483] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] for vif in network_info: [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return self._sync_wrapper(fn, *args, **kwargs) [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.wait() [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self[:] = self._gt.wait() [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return self._exit_event.wait() [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 642.803857] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] current.throw(*self._exc) [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] result = function(*args, **kwargs) [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return func(*args, **kwargs) [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise e [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] nwinfo = self.network_api.allocate_for_instance( [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] created_port_ids = self._update_ports_for_instance( [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] with excutils.save_and_reraise_exception(): [ 642.804232] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.force_reraise() [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise self.value [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] updated_port = self._update_port( [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] _ensure_no_port_binding_failure(port) [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise exception.PortBindingFailed(port_id=port['id']) [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 642.804601] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] [ 642.804601] env[62000]: INFO nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Terminating instance [ 642.806331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.806490] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquired lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.806658] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 643.135491] env[62000]: INFO nova.compute.manager [-] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Took 1.03 seconds to deallocate network for instance. [ 643.137961] env[62000]: DEBUG nova.compute.claims [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 643.138162] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.251404] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.526s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 643.251987] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 643.254644] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.273s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 643.256030] env[62000]: INFO nova.compute.claims [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 643.322995] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 643.410948] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.761050] env[62000]: DEBUG nova.compute.utils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.765238] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.765443] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 643.842516] env[62000]: DEBUG nova.policy [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c822898bf3724a36b6a97a4df1f8513d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d64263849664f2eb4913ec0c7a45949', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.916016] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Releasing lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.916016] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 643.916016] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 643.916016] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e53d00bc-640f-4151-a1eb-6d3a9c34767b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.923906] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb05e8a-d1ee-4027-a77d-f4f3ce461a9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.941417] env[62000]: DEBUG nova.compute.manager [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Received event network-changed-4f33ab6c-e2f2-484d-84e8-44ca61b4ca38 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 643.941714] env[62000]: DEBUG nova.compute.manager [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Refreshing instance network info cache due to event network-changed-4f33ab6c-e2f2-484d-84e8-44ca61b4ca38. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 643.942088] env[62000]: DEBUG oslo_concurrency.lockutils [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] Acquiring lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 643.942316] env[62000]: DEBUG oslo_concurrency.lockutils [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] Acquired lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.942582] env[62000]: DEBUG nova.network.neutron [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Refreshing network info cache for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 643.958374] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8a414b1b-4513-405a-a0f0-a0af61722e3d could not be found. [ 643.958712] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 643.959009] env[62000]: INFO nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 643.959389] env[62000]: DEBUG oslo.service.loopingcall [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 643.959706] env[62000]: DEBUG nova.compute.manager [-] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.959858] env[62000]: DEBUG nova.network.neutron [-] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 643.984658] env[62000]: DEBUG nova.network.neutron [-] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.199430] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Successfully created port: 72dc94bd-f2a1-44b7-8dfa-913af27ae373 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 644.266114] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 644.463620] env[62000]: DEBUG nova.network.neutron [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 644.491108] env[62000]: DEBUG nova.network.neutron [-] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.598084] env[62000]: DEBUG nova.network.neutron [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.652028] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbbdcda-e47b-4122-85ad-b56cfaedb597 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.662587] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9fdbdb-bec5-4985-a7e8-dbc784f0f6f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.692972] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2d6193-8cc0-49f7-affe-d3cf7a2d92ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.699999] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cac0ba07-9af7-4b12-a27f-878ce868cdb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.712951] env[62000]: DEBUG nova.compute.provider_tree [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.994112] env[62000]: INFO nova.compute.manager [-] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Took 1.03 seconds to deallocate network for instance. [ 644.996673] env[62000]: DEBUG nova.compute.claims [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 644.996911] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.103507] env[62000]: DEBUG oslo_concurrency.lockutils [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] Releasing lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.103814] env[62000]: DEBUG nova.compute.manager [req-66894bcb-7aa9-4958-8e42-1246bb8684ba req-f85eaeaa-eea2-4c08-95df-8d1f1b6578cf service nova] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Received event network-vif-deleted-4f33ab6c-e2f2-484d-84e8-44ca61b4ca38 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 645.215882] env[62000]: DEBUG nova.scheduler.client.report [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 645.277548] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 645.307272] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 645.308024] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 645.308024] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 645.308024] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 645.308188] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 645.308188] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 645.308481] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 645.308534] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 645.308694] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 645.308852] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 645.309032] env[62000]: DEBUG nova.virt.hardware [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 645.309897] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49865664-a2b7-4063-bf9e-6ca9a0c4410f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.318975] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e1e18f-ca02-4ea6-8db5-87133588a7ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.361218] env[62000]: ERROR nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 645.361218] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.361218] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.361218] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.361218] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.361218] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.361218] env[62000]: ERROR nova.compute.manager raise self.value [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.361218] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 645.361218] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.361218] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 645.361601] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.361601] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 645.361601] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 645.361601] env[62000]: ERROR nova.compute.manager [ 645.361601] env[62000]: Traceback (most recent call last): [ 645.361601] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 645.361601] env[62000]: listener.cb(fileno) [ 645.361601] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.361601] env[62000]: result = function(*args, **kwargs) [ 645.361601] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.361601] env[62000]: return func(*args, **kwargs) [ 645.361601] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.361601] env[62000]: raise e [ 645.361601] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.361601] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 645.361601] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.361601] env[62000]: created_port_ids = self._update_ports_for_instance( [ 645.361601] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.361601] env[62000]: with excutils.save_and_reraise_exception(): [ 645.361601] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.361601] env[62000]: self.force_reraise() [ 645.361601] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.361601] env[62000]: raise self.value [ 645.361601] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.361601] env[62000]: updated_port = self._update_port( [ 645.361601] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.361601] env[62000]: _ensure_no_port_binding_failure(port) [ 645.361601] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.361601] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 645.362471] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 645.362471] env[62000]: Removing descriptor: 19 [ 645.362471] env[62000]: ERROR nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Traceback (most recent call last): [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] yield resources [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.driver.spawn(context, instance, image_meta, [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 645.362471] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] vm_ref = self.build_virtual_machine(instance, [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] for vif in network_info: [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self._sync_wrapper(fn, *args, **kwargs) [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.wait() [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self[:] = self._gt.wait() [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self._exit_event.wait() [ 645.362801] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] result = hub.switch() [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self.greenlet.switch() [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] result = function(*args, **kwargs) [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return func(*args, **kwargs) [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise e [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] nwinfo = self.network_api.allocate_for_instance( [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 645.363131] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] created_port_ids = self._update_ports_for_instance( [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] with excutils.save_and_reraise_exception(): [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.force_reraise() [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise self.value [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] updated_port = self._update_port( [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] _ensure_no_port_binding_failure(port) [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 645.363503] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise exception.PortBindingFailed(port_id=port['id']) [ 645.363819] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 645.363819] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] [ 645.363819] env[62000]: INFO nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Terminating instance [ 645.364684] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 645.364840] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquired lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 645.365015] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 645.725070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.725331] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.727660] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.775s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.729705] env[62000]: INFO nova.compute.claims [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.887234] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 645.952050] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.961284] env[62000]: DEBUG nova.compute.manager [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Received event network-changed-72dc94bd-f2a1-44b7-8dfa-913af27ae373 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 645.961470] env[62000]: DEBUG nova.compute.manager [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Refreshing instance network info cache due to event network-changed-72dc94bd-f2a1-44b7-8dfa-913af27ae373. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 645.961830] env[62000]: DEBUG oslo_concurrency.lockutils [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] Acquiring lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.234137] env[62000]: DEBUG nova.compute.utils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 646.237473] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 646.237648] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 646.278898] env[62000]: DEBUG nova.policy [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b0014136775f42db92d5dbc4f09906ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9eb068823bc74cc99499568ebd94650d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.456648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Releasing lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 646.457032] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 646.457426] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 646.458819] env[62000]: DEBUG oslo_concurrency.lockutils [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] Acquired lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.458819] env[62000]: DEBUG nova.network.neutron [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Refreshing network info cache for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 646.460716] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa4d2a87-23ee-469c-b50e-3fe783357095 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.475600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df95c1de-7640-4eff-9676-8038c02be76c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.502059] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ef8c2a0-771f-4931-a183-440fd5d998fe could not be found. [ 646.502309] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 646.502490] env[62000]: INFO nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Took 0.05 seconds to destroy the instance on the hypervisor. [ 646.502729] env[62000]: DEBUG oslo.service.loopingcall [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 646.503288] env[62000]: DEBUG nova.compute.manager [-] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 646.503394] env[62000]: DEBUG nova.network.neutron [-] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 646.527392] env[62000]: DEBUG nova.network.neutron [-] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 646.586084] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Successfully created port: 820c6549-d97c-454b-8e7a-3f3fcfa5d09d {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.738620] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.986752] env[62000]: DEBUG nova.network.neutron [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 647.030066] env[62000]: DEBUG nova.network.neutron [-] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.072138] env[62000]: DEBUG nova.network.neutron [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.102995] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c27028c-704b-41de-9a5f-9d907ec47fb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.111561] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f19a892-f146-483c-b41c-cc4703c9545d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.141088] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feffcd76-2f32-41db-a1f0-a505bffaae0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.148260] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de7c8a9a-94d9-4694-bcff-c4a4d6e6235f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.162910] env[62000]: DEBUG nova.compute.provider_tree [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.534581] env[62000]: INFO nova.compute.manager [-] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Took 1.03 seconds to deallocate network for instance. [ 647.537322] env[62000]: DEBUG nova.compute.claims [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.537498] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.574281] env[62000]: DEBUG oslo_concurrency.lockutils [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] Releasing lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.574573] env[62000]: DEBUG nova.compute.manager [req-d652f360-1df4-4318-bd3a-9cf1b37eaaf4 req-7f0c5f01-a6b3-4897-bb97-0e09fc84f97d service nova] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Received event network-vif-deleted-72dc94bd-f2a1-44b7-8dfa-913af27ae373 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 647.613626] env[62000]: ERROR nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 647.613626] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.613626] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.613626] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.613626] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.613626] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.613626] env[62000]: ERROR nova.compute.manager raise self.value [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.613626] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.613626] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.613626] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.614087] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.614087] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.614087] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 647.614087] env[62000]: ERROR nova.compute.manager [ 647.614087] env[62000]: Traceback (most recent call last): [ 647.614087] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.614087] env[62000]: listener.cb(fileno) [ 647.614087] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.614087] env[62000]: result = function(*args, **kwargs) [ 647.614087] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.614087] env[62000]: return func(*args, **kwargs) [ 647.614087] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.614087] env[62000]: raise e [ 647.614087] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.614087] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 647.614087] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.614087] env[62000]: created_port_ids = self._update_ports_for_instance( [ 647.614087] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.614087] env[62000]: with excutils.save_and_reraise_exception(): [ 647.614087] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.614087] env[62000]: self.force_reraise() [ 647.614087] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.614087] env[62000]: raise self.value [ 647.614087] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.614087] env[62000]: updated_port = self._update_port( [ 647.614087] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.614087] env[62000]: _ensure_no_port_binding_failure(port) [ 647.614087] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.614087] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.614878] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 647.614878] env[62000]: Removing descriptor: 19 [ 647.667759] env[62000]: DEBUG nova.scheduler.client.report [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.751686] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.776446] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.776700] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.776854] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.777059] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.777207] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.777350] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.777547] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.777702] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.777863] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.778034] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.778209] env[62000]: DEBUG nova.virt.hardware [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.779103] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d6f7aa-3dfb-40de-afaa-9e81b5970478 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.787837] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b2acd38-2181-4b63-8618-93bab1cedb03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.800615] env[62000]: ERROR nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Traceback (most recent call last): [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] yield resources [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.driver.spawn(context, instance, image_meta, [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] vm_ref = self.build_virtual_machine(instance, [ 647.800615] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] for vif in network_info: [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return self._sync_wrapper(fn, *args, **kwargs) [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.wait() [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self[:] = self._gt.wait() [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return self._exit_event.wait() [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 647.801024] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] current.throw(*self._exc) [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] result = function(*args, **kwargs) [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return func(*args, **kwargs) [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise e [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] nwinfo = self.network_api.allocate_for_instance( [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] created_port_ids = self._update_ports_for_instance( [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] with excutils.save_and_reraise_exception(): [ 647.801358] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.force_reraise() [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise self.value [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] updated_port = self._update_port( [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] _ensure_no_port_binding_failure(port) [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise exception.PortBindingFailed(port_id=port['id']) [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 647.801771] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] [ 647.801771] env[62000]: INFO nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Terminating instance [ 647.803026] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.803202] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquired lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.803369] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 648.018465] env[62000]: DEBUG nova.compute.manager [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Received event network-changed-820c6549-d97c-454b-8e7a-3f3fcfa5d09d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 648.018679] env[62000]: DEBUG nova.compute.manager [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Refreshing instance network info cache due to event network-changed-820c6549-d97c-454b-8e7a-3f3fcfa5d09d. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 648.018811] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] Acquiring lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.172168] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.172708] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 648.175620] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.257s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.175794] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.175949] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 648.176242] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.179577] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d403a255-24ef-43a8-9ab7-65f860caa780 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.187756] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef3f7aa2-e79b-4739-92f8-f2c7c222a737 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.201369] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b892794c-b5f9-437c-8375-5e4371904bf4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.208348] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c544700-8cb7-4fea-834b-1dc5571b7b35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.238530] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181627MB free_disk=53GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 648.238681] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 648.322243] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 648.411795] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.684062] env[62000]: DEBUG nova.compute.utils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.685404] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.687796] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 648.739438] env[62000]: DEBUG nova.policy [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abd23fd5a7bb43d08acd62244d19c153', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4300067ba40e4949861567c49e3f8093', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.914109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Releasing lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.914546] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.914731] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 648.915296] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] Acquired lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.915485] env[62000]: DEBUG nova.network.neutron [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Refreshing network info cache for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 648.916935] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-546108c0-9519-4ae3-b032-33291f2b8ff5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.925974] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f809245-40de-491c-aa84-f1284466d847 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.952560] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6 could not be found. [ 648.952784] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 648.952971] env[62000]: INFO nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.953242] env[62000]: DEBUG oslo.service.loopingcall [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.955842] env[62000]: DEBUG nova.compute.manager [-] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.955949] env[62000]: DEBUG nova.network.neutron [-] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 648.971925] env[62000]: DEBUG nova.network.neutron [-] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.057791] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Successfully created port: db189edf-a178-44ef-ba57-b3ac3d4d4e9c {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.062927] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f91b410f-06a0-4bdb-8e72-75cce0a69914 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.070592] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62c05285-8ecb-41c6-8fa8-74f3ba876ba9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.100915] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3493f40-26dc-4ba2-9198-df598cd97716 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.108109] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f15bf1ef-62e1-460f-a9d8-40a6e49a9945 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.128232] env[62000]: DEBUG nova.compute.provider_tree [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.189508] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 649.436223] env[62000]: DEBUG nova.network.neutron [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 649.474370] env[62000]: DEBUG nova.network.neutron [-] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.550789] env[62000]: DEBUG nova.network.neutron [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.630833] env[62000]: DEBUG nova.scheduler.client.report [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.977493] env[62000]: INFO nova.compute.manager [-] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Took 1.02 seconds to deallocate network for instance. [ 649.979784] env[62000]: DEBUG nova.compute.claims [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.980035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 650.013144] env[62000]: ERROR nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 650.013144] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.013144] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.013144] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.013144] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.013144] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.013144] env[62000]: ERROR nova.compute.manager raise self.value [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.013144] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.013144] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.013144] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.013616] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.013616] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.013616] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 650.013616] env[62000]: ERROR nova.compute.manager [ 650.013616] env[62000]: Traceback (most recent call last): [ 650.013616] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.013616] env[62000]: listener.cb(fileno) [ 650.013616] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.013616] env[62000]: result = function(*args, **kwargs) [ 650.013616] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.013616] env[62000]: return func(*args, **kwargs) [ 650.013616] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.013616] env[62000]: raise e [ 650.013616] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.013616] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 650.013616] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.013616] env[62000]: created_port_ids = self._update_ports_for_instance( [ 650.013616] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.013616] env[62000]: with excutils.save_and_reraise_exception(): [ 650.013616] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.013616] env[62000]: self.force_reraise() [ 650.013616] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.013616] env[62000]: raise self.value [ 650.013616] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.013616] env[62000]: updated_port = self._update_port( [ 650.013616] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.013616] env[62000]: _ensure_no_port_binding_failure(port) [ 650.013616] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.013616] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.014404] env[62000]: nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 650.014404] env[62000]: Removing descriptor: 19 [ 650.041414] env[62000]: DEBUG nova.compute.manager [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Received event network-changed-db189edf-a178-44ef-ba57-b3ac3d4d4e9c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 650.041613] env[62000]: DEBUG nova.compute.manager [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Refreshing instance network info cache due to event network-changed-db189edf-a178-44ef-ba57-b3ac3d4d4e9c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 650.041824] env[62000]: DEBUG oslo_concurrency.lockutils [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] Acquiring lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.041987] env[62000]: DEBUG oslo_concurrency.lockutils [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] Acquired lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.042168] env[62000]: DEBUG nova.network.neutron [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Refreshing network info cache for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 650.053481] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] Releasing lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 650.053696] env[62000]: DEBUG nova.compute.manager [req-aa67e6b1-8876-40cd-9187-74995aa95540 req-b07ef173-3964-4464-ad63-a438927a9e1e service nova] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Received event network-vif-deleted-820c6549-d97c-454b-8e7a-3f3fcfa5d09d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 650.135645] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.959s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.136312] env[62000]: ERROR nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Traceback (most recent call last): [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.driver.spawn(context, instance, image_meta, [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] vm_ref = self.build_virtual_machine(instance, [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.136312] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] for vif in network_info: [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self._sync_wrapper(fn, *args, **kwargs) [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.wait() [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self[:] = self._gt.wait() [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self._exit_event.wait() [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] result = hub.switch() [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.136610] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return self.greenlet.switch() [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] result = function(*args, **kwargs) [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] return func(*args, **kwargs) [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise e [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] nwinfo = self.network_api.allocate_for_instance( [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] created_port_ids = self._update_ports_for_instance( [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] with excutils.save_and_reraise_exception(): [ 650.136914] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] self.force_reraise() [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise self.value [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] updated_port = self._update_port( [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] _ensure_no_port_binding_failure(port) [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] raise exception.PortBindingFailed(port_id=port['id']) [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] nova.exception.PortBindingFailed: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. [ 650.137427] env[62000]: ERROR nova.compute.manager [instance: b6186f74-8a8e-4145-8126-7048ad29673f] [ 650.137934] env[62000]: DEBUG nova.compute.utils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 650.138211] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.574s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.141299] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Build of instance b6186f74-8a8e-4145-8126-7048ad29673f was re-scheduled: Binding failed for port 19a392f0-dc37-4c74-b21e-210b182b9d05, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 650.141735] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 650.141989] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquiring lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.142157] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Acquired lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.142321] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 650.197743] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 650.224829] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 650.225094] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 650.225257] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 650.225439] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 650.225585] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 650.225731] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 650.225933] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 650.226113] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 650.226279] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 650.226437] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 650.226604] env[62000]: DEBUG nova.virt.hardware [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 650.227446] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164a9cec-f6a4-4964-8ca3-0ecf88e26098 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.235548] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd7206e-70b7-4774-bdc2-d290e8b806ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.250021] env[62000]: ERROR nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Traceback (most recent call last): [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] yield resources [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.driver.spawn(context, instance, image_meta, [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] vm_ref = self.build_virtual_machine(instance, [ 650.250021] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] for vif in network_info: [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return self._sync_wrapper(fn, *args, **kwargs) [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.wait() [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self[:] = self._gt.wait() [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return self._exit_event.wait() [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 650.250352] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] current.throw(*self._exc) [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] result = function(*args, **kwargs) [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return func(*args, **kwargs) [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise e [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] nwinfo = self.network_api.allocate_for_instance( [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] created_port_ids = self._update_ports_for_instance( [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] with excutils.save_and_reraise_exception(): [ 650.250648] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.force_reraise() [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise self.value [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] updated_port = self._update_port( [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] _ensure_no_port_binding_failure(port) [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise exception.PortBindingFailed(port_id=port['id']) [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 650.250955] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] [ 650.250955] env[62000]: INFO nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Terminating instance [ 650.252020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.563041] env[62000]: DEBUG nova.network.neutron [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.666530] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 650.669659] env[62000]: DEBUG nova.network.neutron [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.754819] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.992019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd4c054-d38f-4e62-9b3b-22aae3f7ec0e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.999196] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ddde130-de84-4f79-8f62-a3b9a7df6c6b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.027703] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebe9fa61-0b05-4345-9538-e7efbb765e06 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.034889] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0302665-098e-4c32-a435-6008dd6db947 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.047678] env[62000]: DEBUG nova.compute.provider_tree [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.173985] env[62000]: DEBUG oslo_concurrency.lockutils [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] Releasing lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.173985] env[62000]: DEBUG nova.compute.manager [req-12f5819b-1a45-4314-b2a8-6a914e579847 req-2ea9e508-9c51-48f2-9d27-34dd531c84ea service nova] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Received event network-vif-deleted-db189edf-a178-44ef-ba57-b3ac3d4d4e9c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 651.173985] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.173985] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 651.257168] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Releasing lock "refresh_cache-b6186f74-8a8e-4145-8126-7048ad29673f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.257383] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 651.257510] env[62000]: DEBUG nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.257679] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 651.272317] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.550799] env[62000]: DEBUG nova.scheduler.client.report [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 651.699526] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 651.775381] env[62000]: DEBUG nova.network.neutron [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.778409] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.056927] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.058069] env[62000]: ERROR nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Traceback (most recent call last): [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.driver.spawn(context, instance, image_meta, [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] vm_ref = self.build_virtual_machine(instance, [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.058069] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] for vif in network_info: [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self._sync_wrapper(fn, *args, **kwargs) [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.wait() [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self[:] = self._gt.wait() [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self._exit_event.wait() [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] result = hub.switch() [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.058391] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return self.greenlet.switch() [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] result = function(*args, **kwargs) [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] return func(*args, **kwargs) [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise e [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] nwinfo = self.network_api.allocate_for_instance( [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] created_port_ids = self._update_ports_for_instance( [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] with excutils.save_and_reraise_exception(): [ 652.058723] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] self.force_reraise() [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise self.value [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] updated_port = self._update_port( [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] _ensure_no_port_binding_failure(port) [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] raise exception.PortBindingFailed(port_id=port['id']) [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] nova.exception.PortBindingFailed: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. [ 652.059068] env[62000]: ERROR nova.compute.manager [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] [ 652.059347] env[62000]: DEBUG nova.compute.utils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.060162] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.853s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.061583] env[62000]: INFO nova.compute.claims [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.064544] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Build of instance 7396796d-77e5-4364-b889-9cf3cbdad3c9 was re-scheduled: Binding failed for port 5a465687-beb6-4d8f-94cb-00f4abada5e1, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.064921] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.065332] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquiring lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.065332] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Acquired lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.065452] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 652.278702] env[62000]: INFO nova.compute.manager [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] [instance: b6186f74-8a8e-4145-8126-7048ad29673f] Took 1.02 seconds to deallocate network for instance. [ 652.281501] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.281932] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.282174] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 652.282639] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5aa01146-9a08-43b9-9e07-78afa84b6470 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.292108] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ca71ae7-fb27-4ba7-88a0-643e31130a92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.313568] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 042678e8-2028-416a-bc4a-c473c7f073bd could not be found. [ 652.313793] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 652.313972] env[62000]: INFO nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Took 0.03 seconds to destroy the instance on the hypervisor. [ 652.314242] env[62000]: DEBUG oslo.service.loopingcall [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 652.315067] env[62000]: DEBUG nova.compute.manager [-] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 652.315185] env[62000]: DEBUG nova.network.neutron [-] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 652.328997] env[62000]: DEBUG nova.network.neutron [-] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.584206] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 652.654031] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.830911] env[62000]: DEBUG nova.network.neutron [-] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.156959] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Releasing lock "refresh_cache-7396796d-77e5-4364-b889-9cf3cbdad3c9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.157258] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 653.157414] env[62000]: DEBUG nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.157587] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 653.177437] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 653.306475] env[62000]: INFO nova.scheduler.client.report [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Deleted allocations for instance b6186f74-8a8e-4145-8126-7048ad29673f [ 653.332768] env[62000]: INFO nova.compute.manager [-] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Took 1.02 seconds to deallocate network for instance. [ 653.334845] env[62000]: DEBUG nova.compute.claims [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 653.334972] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.430534] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36fa905-710f-41e4-b565-97c484c48757 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.437790] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-496844e8-0d1d-4435-a729-602d925c7647 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.467708] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b03e19-24eb-4b34-85ae-c47f2a9aaca9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.475365] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c42d7ff4-396f-4e31-8792-135bd6255c4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.488687] env[62000]: DEBUG nova.compute.provider_tree [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 653.679447] env[62000]: DEBUG nova.network.neutron [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.815793] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3e4cd179-9846-4d90-bcc0-dbd8c2464e1e tempest-FloatingIPsAssociationNegativeTestJSON-1260778380 tempest-FloatingIPsAssociationNegativeTestJSON-1260778380-project-member] Lock "b6186f74-8a8e-4145-8126-7048ad29673f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 95.530s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.993100] env[62000]: DEBUG nova.scheduler.client.report [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.181655] env[62000]: INFO nova.compute.manager [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] [instance: 7396796d-77e5-4364-b889-9cf3cbdad3c9] Took 1.02 seconds to deallocate network for instance. [ 654.317997] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 654.498480] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 654.499038] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 654.501650] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.421s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 654.844037] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 655.010633] env[62000]: DEBUG nova.compute.utils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.012193] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.012411] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 655.095074] env[62000]: DEBUG nova.policy [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c822898bf3724a36b6a97a4df1f8513d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6d64263849664f2eb4913ec0c7a45949', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.219122] env[62000]: INFO nova.scheduler.client.report [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Deleted allocations for instance 7396796d-77e5-4364-b889-9cf3cbdad3c9 [ 655.416134] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Successfully created port: 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 655.443675] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0d651eb-d6d7-4d4b-8724-488f28d6b9f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.451714] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d52e0c-ddca-490e-8f3b-c4afe68e3f6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.487469] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afb4cace-dd95-4d4f-bc7f-f64bfb44e9ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.497835] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45177785-2849-4583-897a-c0e45762bc0f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.512591] env[62000]: DEBUG nova.compute.provider_tree [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 655.520560] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 655.732027] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67516feb-a4c1-4b9d-945a-61c2b4469bdf tempest-ServerRescueNegativeTestJSON-1351341441 tempest-ServerRescueNegativeTestJSON-1351341441-project-member] Lock "7396796d-77e5-4364-b889-9cf3cbdad3c9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.686s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.015996] env[62000]: DEBUG nova.scheduler.client.report [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.235392] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 656.318608] env[62000]: DEBUG nova.compute.manager [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Received event network-changed-0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 656.319179] env[62000]: DEBUG nova.compute.manager [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Refreshing instance network info cache due to event network-changed-0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 656.319403] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] Acquiring lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.319695] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] Acquired lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.320036] env[62000]: DEBUG nova.network.neutron [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Refreshing network info cache for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 656.439399] env[62000]: ERROR nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 656.439399] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.439399] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.439399] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.439399] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.439399] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.439399] env[62000]: ERROR nova.compute.manager raise self.value [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.439399] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 656.439399] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.439399] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 656.439868] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.439868] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 656.439868] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 656.439868] env[62000]: ERROR nova.compute.manager [ 656.439868] env[62000]: Traceback (most recent call last): [ 656.439868] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 656.439868] env[62000]: listener.cb(fileno) [ 656.439868] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.439868] env[62000]: result = function(*args, **kwargs) [ 656.439868] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.439868] env[62000]: return func(*args, **kwargs) [ 656.439868] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.439868] env[62000]: raise e [ 656.439868] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.439868] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 656.439868] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.439868] env[62000]: created_port_ids = self._update_ports_for_instance( [ 656.439868] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.439868] env[62000]: with excutils.save_and_reraise_exception(): [ 656.439868] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.439868] env[62000]: self.force_reraise() [ 656.439868] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.439868] env[62000]: raise self.value [ 656.439868] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.439868] env[62000]: updated_port = self._update_port( [ 656.439868] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.439868] env[62000]: _ensure_no_port_binding_failure(port) [ 656.439868] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.439868] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 656.440664] env[62000]: nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 656.440664] env[62000]: Removing descriptor: 14 [ 656.525493] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.023s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.525783] env[62000]: ERROR nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Traceback (most recent call last): [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.driver.spawn(context, instance, image_meta, [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] vm_ref = self.build_virtual_machine(instance, [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.525783] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] for vif in network_info: [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return self._sync_wrapper(fn, *args, **kwargs) [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.wait() [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self[:] = self._gt.wait() [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return self._exit_event.wait() [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] current.throw(*self._exc) [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.526099] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] result = function(*args, **kwargs) [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] return func(*args, **kwargs) [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise e [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] nwinfo = self.network_api.allocate_for_instance( [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] created_port_ids = self._update_ports_for_instance( [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] with excutils.save_and_reraise_exception(): [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] self.force_reraise() [ 656.526623] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise self.value [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] updated_port = self._update_port( [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] _ensure_no_port_binding_failure(port) [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] raise exception.PortBindingFailed(port_id=port['id']) [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] nova.exception.PortBindingFailed: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. [ 656.526966] env[62000]: ERROR nova.compute.manager [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] [ 656.526966] env[62000]: DEBUG nova.compute.utils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 656.528032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.984s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.532418] env[62000]: INFO nova.compute.claims [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 656.532796] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Build of instance 289e3403-c90d-40f0-9e18-1d10c0f8aec4 was re-scheduled: Binding failed for port b2de1de4-7221-4d58-b5e6-ae5d1c3c71c2, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 656.533222] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 656.533446] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquiring lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.533589] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Acquired lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 656.533746] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 656.535500] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 656.569634] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 656.570113] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 656.570189] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 656.570329] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 656.570475] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 656.570620] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 656.570966] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 656.571062] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 656.571205] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 656.571571] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 656.571799] env[62000]: DEBUG nova.virt.hardware [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 656.572779] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd0518c-db66-472c-8df5-41841cac1a00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.581603] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6558573e-b6ea-4878-b2d4-b953b603bdc4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.596790] env[62000]: ERROR nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Traceback (most recent call last): [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] yield resources [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.driver.spawn(context, instance, image_meta, [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] vm_ref = self.build_virtual_machine(instance, [ 656.596790] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] for vif in network_info: [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return self._sync_wrapper(fn, *args, **kwargs) [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.wait() [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self[:] = self._gt.wait() [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return self._exit_event.wait() [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 656.597334] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] current.throw(*self._exc) [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] result = function(*args, **kwargs) [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return func(*args, **kwargs) [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise e [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] nwinfo = self.network_api.allocate_for_instance( [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] created_port_ids = self._update_ports_for_instance( [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] with excutils.save_and_reraise_exception(): [ 656.597837] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.force_reraise() [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise self.value [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] updated_port = self._update_port( [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] _ensure_no_port_binding_failure(port) [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise exception.PortBindingFailed(port_id=port['id']) [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 656.598449] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] [ 656.598449] env[62000]: INFO nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Terminating instance [ 656.602229] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 656.755712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.840201] env[62000]: DEBUG nova.network.neutron [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 656.908897] env[62000]: DEBUG nova.network.neutron [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.066609] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.146060] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.411304] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c4d9865-66bf-43a0-8d59-be1dc5bde7f1 req-c466ba47-9f69-4763-bbe4-7fae28ade24d service nova] Releasing lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.411848] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquired lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.411934] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 657.650896] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Releasing lock "refresh_cache-289e3403-c90d-40f0-9e18-1d10c0f8aec4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 657.651102] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 657.651294] env[62000]: DEBUG nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 657.651463] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 657.676438] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 657.938743] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.017281] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.035947] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe0911c-4b60-485d-a4e6-5f95d0ae933c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.044483] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11fe485-be56-43b7-b747-99d9730577f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.078031] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d9b3ef5-ff2d-4574-a290-c1de882837c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.086127] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb983f2-7e00-43dc-802a-eabbd24a4118 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.099111] env[62000]: DEBUG nova.compute.provider_tree [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.179545] env[62000]: DEBUG nova.network.neutron [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.346651] env[62000]: DEBUG nova.compute.manager [req-2293e248-757b-4366-bd0c-774db0ba4e74 req-d5b50d95-5895-4a8a-bb1e-506ec7391ae6 service nova] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Received event network-vif-deleted-0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 658.519448] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Releasing lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.519885] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 658.520110] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 658.520407] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-70546af9-4c87-40c2-962a-cf9438b495ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.529099] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7827236-c144-49cd-87a2-40329e90ff20 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.552574] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4 could not be found. [ 658.552906] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 658.553136] env[62000]: INFO nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 658.553396] env[62000]: DEBUG oslo.service.loopingcall [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 658.553597] env[62000]: DEBUG nova.compute.manager [-] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.553689] env[62000]: DEBUG nova.network.neutron [-] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 658.568307] env[62000]: DEBUG nova.network.neutron [-] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 658.602047] env[62000]: DEBUG nova.scheduler.client.report [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.681947] env[62000]: INFO nova.compute.manager [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] [instance: 289e3403-c90d-40f0-9e18-1d10c0f8aec4] Took 1.03 seconds to deallocate network for instance. [ 659.071597] env[62000]: DEBUG nova.network.neutron [-] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.107683] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.579s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.107683] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.110306] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.972s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.574571] env[62000]: INFO nova.compute.manager [-] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Took 1.02 seconds to deallocate network for instance. [ 659.576913] env[62000]: DEBUG nova.compute.claims [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 659.577109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.614610] env[62000]: DEBUG nova.compute.utils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 659.618831] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 659.619009] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 659.671249] env[62000]: DEBUG nova.policy [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abd23fd5a7bb43d08acd62244d19c153', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4300067ba40e4949861567c49e3f8093', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 659.707097] env[62000]: INFO nova.scheduler.client.report [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Deleted allocations for instance 289e3403-c90d-40f0-9e18-1d10c0f8aec4 [ 659.991032] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db0d940f-de16-4613-9628-522c73433219 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.000144] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7334fbec-895f-4f06-b01f-1b626a33fc8f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.036408] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Successfully created port: adfb24f7-ba06-4faf-bb95-ebfcbc767b3f {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.039067] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7ba96a-b688-45c7-b5df-fe90f28864b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.046437] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69679f32-f36f-46e1-aa99-1a95278b681f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.059907] env[62000]: DEBUG nova.compute.provider_tree [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.120685] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.217745] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d626e5a3-c18b-4000-82f0-509f4e826034 tempest-ServerActionsTestJSON-1792958832 tempest-ServerActionsTestJSON-1792958832-project-member] Lock "289e3403-c90d-40f0-9e18-1d10c0f8aec4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.563s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 660.563486] env[62000]: DEBUG nova.scheduler.client.report [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 660.720489] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 660.841186] env[62000]: DEBUG nova.compute.manager [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Received event network-changed-adfb24f7-ba06-4faf-bb95-ebfcbc767b3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 660.841186] env[62000]: DEBUG nova.compute.manager [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Refreshing instance network info cache due to event network-changed-adfb24f7-ba06-4faf-bb95-ebfcbc767b3f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 660.841186] env[62000]: DEBUG oslo_concurrency.lockutils [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] Acquiring lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 660.841186] env[62000]: DEBUG oslo_concurrency.lockutils [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] Acquired lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.841186] env[62000]: DEBUG nova.network.neutron [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Refreshing network info cache for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 661.068819] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.069535] env[62000]: ERROR nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Traceback (most recent call last): [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.driver.spawn(context, instance, image_meta, [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] vm_ref = self.build_virtual_machine(instance, [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.069535] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] for vif in network_info: [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return self._sync_wrapper(fn, *args, **kwargs) [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.wait() [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self[:] = self._gt.wait() [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return self._exit_event.wait() [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] current.throw(*self._exc) [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.069888] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] result = function(*args, **kwargs) [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] return func(*args, **kwargs) [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise e [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] nwinfo = self.network_api.allocate_for_instance( [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] created_port_ids = self._update_ports_for_instance( [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] with excutils.save_and_reraise_exception(): [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] self.force_reraise() [ 661.070207] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise self.value [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] updated_port = self._update_port( [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] _ensure_no_port_binding_failure(port) [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] raise exception.PortBindingFailed(port_id=port['id']) [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] nova.exception.PortBindingFailed: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. [ 661.070511] env[62000]: ERROR nova.compute.manager [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] [ 661.074730] env[62000]: DEBUG nova.compute.utils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 661.077321] env[62000]: ERROR nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 661.077321] env[62000]: ERROR nova.compute.manager Traceback (most recent call last): [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.077321] env[62000]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.077321] env[62000]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.077321] env[62000]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.077321] env[62000]: ERROR nova.compute.manager self.force_reraise() [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.077321] env[62000]: ERROR nova.compute.manager raise self.value [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.077321] env[62000]: ERROR nova.compute.manager updated_port = self._update_port( [ 661.077321] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.077321] env[62000]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 661.077722] env[62000]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.077722] env[62000]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 661.077722] env[62000]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 661.077722] env[62000]: ERROR nova.compute.manager [ 661.078101] env[62000]: Traceback (most recent call last): [ 661.078154] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 661.078154] env[62000]: listener.cb(fileno) [ 661.078154] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.078154] env[62000]: result = function(*args, **kwargs) [ 661.078154] env[62000]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.078154] env[62000]: return func(*args, **kwargs) [ 661.078154] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.078154] env[62000]: raise e [ 661.078154] env[62000]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.078154] env[62000]: nwinfo = self.network_api.allocate_for_instance( [ 661.078154] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.078154] env[62000]: created_port_ids = self._update_ports_for_instance( [ 661.078154] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.078154] env[62000]: with excutils.save_and_reraise_exception(): [ 661.078533] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.078533] env[62000]: self.force_reraise() [ 661.078533] env[62000]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.078533] env[62000]: raise self.value [ 661.078533] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.078533] env[62000]: updated_port = self._update_port( [ 661.078533] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.078533] env[62000]: _ensure_no_port_binding_failure(port) [ 661.078533] env[62000]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.078533] env[62000]: raise exception.PortBindingFailed(port_id=port['id']) [ 661.078533] env[62000]: nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 661.078533] env[62000]: Removing descriptor: 14 [ 661.078824] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.082s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.082335] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Build of instance ab167fa4-062b-4137-801c-9b57a17b20e2 was re-scheduled: Binding failed for port ee1823ff-c772-4e4f-ba45-7b5ec4dde1df, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 661.082821] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 661.083246] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquiring lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.083380] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Acquired lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.083988] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 661.135435] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.163656] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.163910] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.164077] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.164264] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.164409] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.164556] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.164758] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.164914] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.165094] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.165253] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.165425] env[62000]: DEBUG nova.virt.hardware [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.166812] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-199e7b72-1d10-4dad-a40d-06b9891c348b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.175134] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6d30c78-c08b-4a48-9aed-5a5c16b0b83e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.190098] env[62000]: ERROR nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Traceback (most recent call last): [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] yield resources [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.driver.spawn(context, instance, image_meta, [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] vm_ref = self.build_virtual_machine(instance, [ 661.190098] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] for vif in network_info: [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return self._sync_wrapper(fn, *args, **kwargs) [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.wait() [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self[:] = self._gt.wait() [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return self._exit_event.wait() [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 661.190439] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] current.throw(*self._exc) [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] result = function(*args, **kwargs) [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return func(*args, **kwargs) [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise e [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] nwinfo = self.network_api.allocate_for_instance( [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] created_port_ids = self._update_ports_for_instance( [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] with excutils.save_and_reraise_exception(): [ 661.190741] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.force_reraise() [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise self.value [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] updated_port = self._update_port( [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] _ensure_no_port_binding_failure(port) [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise exception.PortBindingFailed(port_id=port['id']) [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 661.191070] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] [ 661.191070] env[62000]: INFO nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Terminating instance [ 661.192283] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.242361] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.357871] env[62000]: DEBUG nova.network.neutron [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.446433] env[62000]: DEBUG nova.network.neutron [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.617626] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 661.691989] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.950593] env[62000]: DEBUG oslo_concurrency.lockutils [req-29f11328-33c3-4121-9a5c-d26501b60ef9 req-ca145bf6-d7f9-4419-9099-e9dc8ef71f32 service nova] Releasing lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.951824] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.951824] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 662.026151] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7c78c2b-5b81-481c-a29c-61dbc724da30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.034587] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef531bc-1095-414d-80a4-2d0b0436aa44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.070598] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c4cd93a-6d3b-418f-b4e3-89825f846f91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.078813] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a03443f-09b8-4b16-8282-f5c20154f10c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.091973] env[62000]: DEBUG nova.compute.provider_tree [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 662.193747] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Releasing lock "refresh_cache-ab167fa4-062b-4137-801c-9b57a17b20e2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.194038] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 662.194472] env[62000]: DEBUG nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 662.194766] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 662.211064] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.471439] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 662.563819] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.594601] env[62000]: DEBUG nova.scheduler.client.report [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 662.714300] env[62000]: DEBUG nova.network.neutron [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.867117] env[62000]: DEBUG nova.compute.manager [req-4a1aa5aa-b3a2-4a15-af2d-218b16a5a8e3 req-3ed0d63b-f33c-4c7f-9cdd-c29f18602760 service nova] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Received event network-vif-deleted-adfb24f7-ba06-4faf-bb95-ebfcbc767b3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 663.072129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.075020] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.075020] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 663.075020] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8dbd2ca-e223-4867-acf9-4ef054ebdf5c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.083996] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45d305a4-53b9-4d98-aa8e-956db5d51358 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.099165] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 663.100174] env[62000]: ERROR nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Traceback (most recent call last): [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.driver.spawn(context, instance, image_meta, [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] vm_ref = self.build_virtual_machine(instance, [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] vif_infos = vmwarevif.get_vif_info(self._session, [ 663.100174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] for vif in network_info: [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return self._sync_wrapper(fn, *args, **kwargs) [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.wait() [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self[:] = self._gt.wait() [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return self._exit_event.wait() [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] current.throw(*self._exc) [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 663.100459] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] result = function(*args, **kwargs) [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] return func(*args, **kwargs) [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise e [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] nwinfo = self.network_api.allocate_for_instance( [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] created_port_ids = self._update_ports_for_instance( [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] with excutils.save_and_reraise_exception(): [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] self.force_reraise() [ 663.100851] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise self.value [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] updated_port = self._update_port( [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] _ensure_no_port_binding_failure(port) [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] raise exception.PortBindingFailed(port_id=port['id']) [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] nova.exception.PortBindingFailed: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. [ 663.101174] env[62000]: ERROR nova.compute.manager [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] [ 663.101174] env[62000]: DEBUG nova.compute.utils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 663.102543] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.565s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.109930] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Build of instance 8a414b1b-4513-405a-a0f0-a0af61722e3d was re-scheduled: Binding failed for port 4f33ab6c-e2f2-484d-84e8-44ca61b4ca38, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 663.111852] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 663.111852] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.111852] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquired lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.111852] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 663.112219] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f36c4c0a-6777-468f-8798-44093023b2d1 could not be found. [ 663.112355] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 663.112531] env[62000]: INFO nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Took 0.04 seconds to destroy the instance on the hypervisor. [ 663.112767] env[62000]: DEBUG oslo.service.loopingcall [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.113326] env[62000]: DEBUG nova.compute.manager [-] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.113420] env[62000]: DEBUG nova.network.neutron [-] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 663.134071] env[62000]: DEBUG nova.network.neutron [-] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.216872] env[62000]: INFO nova.compute.manager [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] [instance: ab167fa4-062b-4137-801c-9b57a17b20e2] Took 1.02 seconds to deallocate network for instance. [ 663.641218] env[62000]: DEBUG nova.network.neutron [-] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.648895] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 663.735087] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.933090] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a01bcf8e-2c88-4a3f-83b2-d753dcd54498 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.940755] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c3d7ea-bdaf-421e-80e1-931b3b4350b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.971982] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238a1b06-dbb8-4b1d-9ce3-c5382f848d19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.979393] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e06bd6-4e0a-4666-a908-d797be123582 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.992508] env[62000]: DEBUG nova.compute.provider_tree [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 664.054622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.054867] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.146113] env[62000]: INFO nova.compute.manager [-] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Took 1.03 seconds to deallocate network for instance. [ 664.148526] env[62000]: DEBUG nova.compute.claims [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Aborting claim: {{(pid=62000) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 664.148701] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.241968] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Releasing lock "refresh_cache-8a414b1b-4513-405a-a0f0-a0af61722e3d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.242259] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 664.242446] env[62000]: DEBUG nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.242625] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 664.265272] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 664.274915] env[62000]: INFO nova.scheduler.client.report [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Deleted allocations for instance ab167fa4-062b-4137-801c-9b57a17b20e2 [ 664.498025] env[62000]: DEBUG nova.scheduler.client.report [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 664.769875] env[62000]: DEBUG nova.network.neutron [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.782149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bff4e30b-37df-4dd2-be78-55cdb2855e8c tempest-DeleteServersAdminTestJSON-813093497 tempest-DeleteServersAdminTestJSON-813093497-project-member] Lock "ab167fa4-062b-4137-801c-9b57a17b20e2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.128s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.001092] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.898s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 665.001774] env[62000]: ERROR nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Traceback (most recent call last): [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.driver.spawn(context, instance, image_meta, [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self._vmops.spawn(context, instance, image_meta, injected_files, [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] vm_ref = self.build_virtual_machine(instance, [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] vif_infos = vmwarevif.get_vif_info(self._session, [ 665.001774] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] for vif in network_info: [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self._sync_wrapper(fn, *args, **kwargs) [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.wait() [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self[:] = self._gt.wait() [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self._exit_event.wait() [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] result = hub.switch() [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 665.002145] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return self.greenlet.switch() [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] result = function(*args, **kwargs) [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] return func(*args, **kwargs) [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise e [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] nwinfo = self.network_api.allocate_for_instance( [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] created_port_ids = self._update_ports_for_instance( [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] with excutils.save_and_reraise_exception(): [ 665.002505] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] self.force_reraise() [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise self.value [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] updated_port = self._update_port( [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] _ensure_no_port_binding_failure(port) [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] raise exception.PortBindingFailed(port_id=port['id']) [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] nova.exception.PortBindingFailed: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. [ 665.002894] env[62000]: ERROR nova.compute.manager [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] [ 665.003224] env[62000]: DEBUG nova.compute.utils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 665.003773] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 16.765s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 665.005389] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Build of instance 5ef8c2a0-771f-4931-a183-440fd5d998fe was re-scheduled: Binding failed for port 72dc94bd-f2a1-44b7-8dfa-913af27ae373, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 665.005800] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 665.006034] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 665.006183] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquired lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 665.006342] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 665.275727] env[62000]: INFO nova.compute.manager [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: 8a414b1b-4513-405a-a0f0-a0af61722e3d] Took 1.03 seconds to deallocate network for instance. [ 665.286018] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 665.527115] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 665.594503] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.806086] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.097230] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Releasing lock "refresh_cache-5ef8c2a0-771f-4931-a183-440fd5d998fe" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 666.097476] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 666.097641] env[62000]: DEBUG nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.097804] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 666.116152] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 666.306541] env[62000]: INFO nova.scheduler.client.report [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Deleted allocations for instance 8a414b1b-4513-405a-a0f0-a0af61722e3d [ 666.541530] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 8a414b1b-4513-405a-a0f0-a0af61722e3d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.619023] env[62000]: DEBUG nova.network.neutron [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.816635] env[62000]: DEBUG oslo_concurrency.lockutils [None req-980fa4d9-cad1-42b1-a278-b36c5ddaa921 tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "8a414b1b-4513-405a-a0f0-a0af61722e3d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.610s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 667.045049] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 5ef8c2a0-771f-4931-a183-440fd5d998fe has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.045049] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 667.045301] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 042678e8-2028-416a-bc4a-c473c7f073bd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 667.045301] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 667.045450] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f36c4c0a-6777-468f-8798-44093023b2d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 667.121725] env[62000]: INFO nova.compute.manager [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 5ef8c2a0-771f-4931-a183-440fd5d998fe] Took 1.02 seconds to deallocate network for instance. [ 667.319307] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 667.549130] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 1419e5a8-a78f-44f1-bfc4-00fcf176bf35 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.842893] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.052612] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 1298037f-9cb7-4e2b-b70f-2ab24efe9b91 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.151227] env[62000]: INFO nova.scheduler.client.report [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Deleted allocations for instance 5ef8c2a0-771f-4931-a183-440fd5d998fe [ 668.556752] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 762e429f-f690-43c4-95eb-877caf1cdad7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.659464] env[62000]: DEBUG oslo_concurrency.lockutils [None req-165ce00e-8eca-4696-b61e-8a97040d9b08 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "5ef8c2a0-771f-4931-a183-440fd5d998fe" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.987s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.058584] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 1d9b67b5-c0cf-41ff-a838-79caf8789609 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.162964] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 669.562031] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 5e80a5c8-030c-4ad8-90c4-26136fa39d71 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.698476] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.065253] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance eb27703f-b657-423a-90a9-a7c024a2e473 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.568954] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 49f6c198-13b7-4c07-81d8-c010e7b0598e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.076039] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 415cc4a5-7610-4678-971d-cd00a0e8b54d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 671.578850] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.083045] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 72688992-333f-459d-9d05-f7c728961a6d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 672.329730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.330016] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.586444] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.089552] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 334fd514-c7b5-402f-935e-1d95f9b5dbc8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 673.592920] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 210af329-4cdb-4c3f-9e82-e72a2ea79421 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.097027] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 596af5ab-1791-42ce-93d2-3e4f0a47dfa2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 674.601331] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 4e617bb5-84e4-4505-99e3-61289826f511 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.104441] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance af49714d-8e50-4159-96a5-cf8f70580471 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 675.608866] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance a9783a2a-a7d3-4fbe-9590-d530ac24fa82 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.113070] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 676.615689] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 71192360-6ee6-4876-bf37-da987a09cbb2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.119768] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f3a6527c-777e-4a07-9482-598de15d4eb3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.622511] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 30d82c70-1401-4a1a-a88b-f798f8fbf96a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 677.622789] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 677.622925] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 677.921740] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a76bf575-1b5b-436f-8416-45824e3fdc16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.929455] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56300ec3-291e-460a-89a3-a6da4fb43a19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.958901] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edff2f8c-8f81-42bb-8187-8bc9832b126b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.966601] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb07c3c7-4bc0-4739-9d42-205d8b5a86c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.978752] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.481962] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 678.986856] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 678.987135] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 13.983s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.987419] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.007s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.747951] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f207fd-8be9-40fe-98a2-9b811909c0af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.755414] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b20e2b6-37f8-4238-bf93-21ccbcaf431d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.786741] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff0afef-3d6c-4232-b2d3-01febb68d917 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.794842] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6b815a7-cb66-4eed-af5c-bbd9c434df50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.808594] env[62000]: DEBUG nova.compute.provider_tree [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 680.313057] env[62000]: DEBUG nova.scheduler.client.report [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 680.818470] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.831s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.818864] env[62000]: ERROR nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Traceback (most recent call last): [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.driver.spawn(context, instance, image_meta, [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] vm_ref = self.build_virtual_machine(instance, [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] vif_infos = vmwarevif.get_vif_info(self._session, [ 680.818864] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] for vif in network_info: [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return self._sync_wrapper(fn, *args, **kwargs) [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.wait() [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self[:] = self._gt.wait() [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return self._exit_event.wait() [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] current.throw(*self._exc) [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 680.819244] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] result = function(*args, **kwargs) [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] return func(*args, **kwargs) [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise e [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] nwinfo = self.network_api.allocate_for_instance( [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] created_port_ids = self._update_ports_for_instance( [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] with excutils.save_and_reraise_exception(): [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] self.force_reraise() [ 680.819610] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise self.value [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] updated_port = self._update_port( [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] _ensure_no_port_binding_failure(port) [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] raise exception.PortBindingFailed(port_id=port['id']) [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] nova.exception.PortBindingFailed: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. [ 680.820032] env[62000]: ERROR nova.compute.manager [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] [ 680.820032] env[62000]: DEBUG nova.compute.utils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.820909] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.486s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.824118] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Build of instance e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6 was re-scheduled: Binding failed for port 820c6549-d97c-454b-8e7a-3f3fcfa5d09d, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 680.824895] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 680.824895] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquiring lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.824895] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Acquired lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.825071] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 681.347216] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.445197] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.622428] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53ce2e6b-e7c2-45da-8a75-710ef8c44dcb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.630058] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bdd850c-437f-4320-be73-e0385510c62f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.658757] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a5ab35-0cb9-4e69-bb87-ca3a08f3b762 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.665774] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-573bcf0d-c493-40e7-9b37-c99c246e8056 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.679634] env[62000]: DEBUG nova.compute.provider_tree [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.947514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Releasing lock "refresh_cache-e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.947750] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 681.947928] env[62000]: DEBUG nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 681.948104] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.962643] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 682.183217] env[62000]: DEBUG nova.scheduler.client.report [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 682.465495] env[62000]: DEBUG nova.network.neutron [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.687580] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.688230] env[62000]: ERROR nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Traceback (most recent call last): [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.driver.spawn(context, instance, image_meta, [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] vm_ref = self.build_virtual_machine(instance, [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.688230] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] for vif in network_info: [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return self._sync_wrapper(fn, *args, **kwargs) [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.wait() [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self[:] = self._gt.wait() [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return self._exit_event.wait() [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] current.throw(*self._exc) [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.688584] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] result = function(*args, **kwargs) [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] return func(*args, **kwargs) [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise e [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] nwinfo = self.network_api.allocate_for_instance( [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] created_port_ids = self._update_ports_for_instance( [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] with excutils.save_and_reraise_exception(): [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] self.force_reraise() [ 682.688951] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise self.value [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] updated_port = self._update_port( [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] _ensure_no_port_binding_failure(port) [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] raise exception.PortBindingFailed(port_id=port['id']) [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] nova.exception.PortBindingFailed: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. [ 682.689330] env[62000]: ERROR nova.compute.manager [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] [ 682.689330] env[62000]: DEBUG nova.compute.utils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.690168] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.846s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.691588] env[62000]: INFO nova.compute.claims [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 682.695237] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Build of instance 042678e8-2028-416a-bc4a-c473c7f073bd was re-scheduled: Binding failed for port db189edf-a178-44ef-ba57-b3ac3d4d4e9c, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 682.695674] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 682.695894] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.696054] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.696218] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.969223] env[62000]: INFO nova.compute.manager [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] [instance: e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6] Took 1.02 seconds to deallocate network for instance. [ 683.214473] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.311898] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.814492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "refresh_cache-042678e8-2028-416a-bc4a-c473c7f073bd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.814761] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 683.814906] env[62000]: DEBUG nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 683.815090] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.832220] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.992294] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1881ab08-c136-405e-bbdf-a7189ccda7d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.998074] env[62000]: INFO nova.scheduler.client.report [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Deleted allocations for instance e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6 [ 684.007408] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7678354-5973-4cd4-b268-d2c9774a03d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.039572] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494f08e1-60a0-4de1-850e-fa737d710c3b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.047757] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-477b722d-0005-4789-86df-9f04edd033af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.061427] env[62000]: DEBUG nova.compute.provider_tree [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 684.334367] env[62000]: DEBUG nova.network.neutron [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.511331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a3699d7-bd00-43f9-8d04-8675dfa9a8ac tempest-ServersAdminTestJSON-1571498747 tempest-ServersAdminTestJSON-1571498747-project-member] Lock "e3c65b37-a2b9-43dd-bcc5-f7aaf1e8d6f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.749s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.564103] env[62000]: DEBUG nova.scheduler.client.report [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 684.837970] env[62000]: INFO nova.compute.manager [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 042678e8-2028-416a-bc4a-c473c7f073bd] Took 1.02 seconds to deallocate network for instance. [ 685.014685] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 685.071020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.071020] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 685.073257] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.318s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.078246] env[62000]: INFO nova.compute.claims [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 685.540258] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.584804] env[62000]: DEBUG nova.compute.utils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 685.589111] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 685.589111] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 685.683854] env[62000]: DEBUG nova.policy [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'abd23fd5a7bb43d08acd62244d19c153', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4300067ba40e4949861567c49e3f8093', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 685.868584] env[62000]: INFO nova.scheduler.client.report [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Deleted allocations for instance 042678e8-2028-416a-bc4a-c473c7f073bd [ 686.089614] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 686.225926] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Successfully created port: 46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 686.378280] env[62000]: DEBUG oslo_concurrency.lockutils [None req-99369157-b8b9-4e3d-886b-b96cc1e79b3c tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "042678e8-2028-416a-bc4a-c473c7f073bd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.408s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.439007] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4603e8-2aa4-46b2-b4f8-bda489abbfca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.446721] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b1a7fd9-4b1d-47c5-a670-4946bf7b9df1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.476628] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16812cbc-60a9-445a-a7d5-51d902c2190e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.486265] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc72c567-a5b6-4fd9-9ee5-c59967b1a1c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.497973] env[62000]: DEBUG nova.compute.provider_tree [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 686.883775] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 687.001362] env[62000]: DEBUG nova.scheduler.client.report [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.100212] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 687.125299] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 687.125569] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 687.125726] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 687.125998] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 687.126185] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 687.126334] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 687.126548] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 687.126702] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 687.126865] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 687.127035] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 687.127212] env[62000]: DEBUG nova.virt.hardware [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 687.128116] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b59063-47aa-46f0-83fb-4ea274c26187 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.137053] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd0904c-bd94-4f00-b8ea-46fd72f9f6b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.404083] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.508488] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.508488] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 687.511168] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.934s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.020704] env[62000]: DEBUG nova.compute.utils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 688.020704] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 688.021418] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 688.115325] env[62000]: DEBUG nova.policy [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '63e13861aad6446bbf266ba67aa293e1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '076d68ba0f9d4cdbbfbdeb879a222c08', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 688.221877] env[62000]: DEBUG nova.compute.manager [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Received event network-vif-plugged-46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 688.222144] env[62000]: DEBUG oslo_concurrency.lockutils [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] Acquiring lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.222356] env[62000]: DEBUG oslo_concurrency.lockutils [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.222520] env[62000]: DEBUG oslo_concurrency.lockutils [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.222690] env[62000]: DEBUG nova.compute.manager [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] No waiting events found dispatching network-vif-plugged-46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 688.222853] env[62000]: WARNING nova.compute.manager [req-537b1d0d-60b0-4d32-9c16-64a6171a1f48 req-51f65a4c-5b36-4fa6-9c2e-9561082e58be service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Received unexpected event network-vif-plugged-46a0ba05-811c-4e47-ad7c-0b6021236d10 for instance with vm_state building and task_state spawning. [ 688.394093] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f183ad5a-375c-454f-968e-9bc101b1a5ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.402598] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb81f0c-3b32-4442-988c-78ec6b25dad5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.437994] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f4e729-6abe-4688-9779-c1a8ea8302e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.446155] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33fe247c-37b3-4910-80ae-5e56f20d8955 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.467524] env[62000]: DEBUG nova.compute.provider_tree [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.467524] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Successfully updated port: 46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 688.533383] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 688.629658] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Successfully created port: ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 688.969125] env[62000]: DEBUG nova.scheduler.client.report [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.973080] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.973177] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.973330] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.480637] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.970s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.481285] env[62000]: ERROR nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Traceback (most recent call last): [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.driver.spawn(context, instance, image_meta, [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] vm_ref = self.build_virtual_machine(instance, [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.481285] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] for vif in network_info: [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return self._sync_wrapper(fn, *args, **kwargs) [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.wait() [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self[:] = self._gt.wait() [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return self._exit_event.wait() [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] current.throw(*self._exc) [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.481600] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] result = function(*args, **kwargs) [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] return func(*args, **kwargs) [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise e [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] nwinfo = self.network_api.allocate_for_instance( [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] created_port_ids = self._update_ports_for_instance( [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] with excutils.save_and_reraise_exception(): [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] self.force_reraise() [ 689.481969] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise self.value [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] updated_port = self._update_port( [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] _ensure_no_port_binding_failure(port) [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] raise exception.PortBindingFailed(port_id=port['id']) [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] nova.exception.PortBindingFailed: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. [ 689.482327] env[62000]: ERROR nova.compute.manager [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] [ 689.482327] env[62000]: DEBUG nova.compute.utils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.485431] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.243s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.486864] env[62000]: INFO nova.compute.claims [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.489783] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Build of instance 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4 was re-scheduled: Binding failed for port 0f5fddc9-0bc2-4fdd-b6ce-d0cc8435a181, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 689.489918] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 689.490096] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquiring lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.490249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Acquired lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.490410] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.520474] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.542526] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 689.570453] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 689.570696] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 689.570851] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 689.571049] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 689.571200] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 689.571345] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 689.571562] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 689.571812] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 689.571931] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 689.572092] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 689.572948] env[62000]: DEBUG nova.virt.hardware [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 689.573183] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9167916-bef6-4cd7-9359-0490571aea7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.581439] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44cc207d-fada-4502-a42c-10a44b69de5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.719550] env[62000]: DEBUG nova.network.neutron [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Updating instance_info_cache with network_info: [{"id": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "address": "fa:16:3e:8e:e8:bf", "network": {"id": "c0a7cb0d-5e6f-4f14-81c5-ccab76f340e2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-151434811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4300067ba40e4949861567c49e3f8093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46a0ba05-81", "ovs_interfaceid": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.014562] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.077341] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.222397] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.222726] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Instance network_info: |[{"id": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "address": "fa:16:3e:8e:e8:bf", "network": {"id": "c0a7cb0d-5e6f-4f14-81c5-ccab76f340e2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-151434811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4300067ba40e4949861567c49e3f8093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46a0ba05-81", "ovs_interfaceid": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 690.223260] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:e8:bf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '796c7fcb-00fd-4692-a44b-7ec550201e86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '46a0ba05-811c-4e47-ad7c-0b6021236d10', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 690.234025] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Creating folder: Project (4300067ba40e4949861567c49e3f8093). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 690.234383] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-964f50b2-f497-4fbe-aa30-5fe73df92684 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.247760] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Created folder: Project (4300067ba40e4949861567c49e3f8093) in parent group-v201431. [ 690.248250] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Creating folder: Instances. Parent ref: group-v201441. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 690.248635] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b8ee7970-d1e8-4807-a67f-1fee5671ad67 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.258668] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Created folder: Instances in parent group-v201441. [ 690.258909] env[62000]: DEBUG oslo.service.loopingcall [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.259149] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 690.259314] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-92d7f1f8-d7f8-4630-ba5a-9978c2d2149c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.282457] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 690.282457] env[62000]: value = "task-881852" [ 690.282457] env[62000]: _type = "Task" [ 690.282457] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.290887] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881852, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.448294] env[62000]: DEBUG nova.compute.manager [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Received event network-changed-46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 690.448294] env[62000]: DEBUG nova.compute.manager [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Refreshing instance network info cache due to event network-changed-46a0ba05-811c-4e47-ad7c-0b6021236d10. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 690.448294] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] Acquiring lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.448294] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] Acquired lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.448643] env[62000]: DEBUG nova.network.neutron [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Refreshing network info cache for port 46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 690.465833] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Successfully updated port: ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 690.580520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Releasing lock "refresh_cache-7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.580799] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 690.580911] env[62000]: DEBUG nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.581164] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.598660] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.789610] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80150275-1144-4b1e-915e-3a1d64caf7d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.799734] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-226abe63-a8d9-4f5a-8781-aada74b57f58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.802706] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881852, 'name': CreateVM_Task, 'duration_secs': 0.294412} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.802897] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 690.836055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.836055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.836055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 690.836055] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574c0a5f-7994-47ea-81dd-46b403633f16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.836055] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-562dc7a9-be96-453d-97af-2aec81218fab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.843107] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-609f7b8f-54ad-4781-b677-b0aabd47a2b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.846771] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 690.846771] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e20af7-620e-3358-60bc-e74df96c6f1f" [ 690.846771] env[62000]: _type = "Task" [ 690.846771] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.863879] env[62000]: DEBUG nova.compute.provider_tree [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.869833] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e20af7-620e-3358-60bc-e74df96c6f1f, 'name': SearchDatastore_Task, 'duration_secs': 0.010747} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 690.870140] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.870458] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 690.870700] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.870842] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.871037] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 690.871265] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c8e05514-0f58-4342-8e7d-b145acdce6b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.879093] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 690.879275] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 690.879973] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9c20db9-c6a3-4643-9fdf-1ccd2f3ac329 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.885398] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 690.885398] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d43ff8-8bf3-952b-b2c3-b1af52850904" [ 690.885398] env[62000]: _type = "Task" [ 690.885398] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 690.892760] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d43ff8-8bf3-952b-b2c3-b1af52850904, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 690.967882] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.968062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquired lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.968218] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.100908] env[62000]: DEBUG nova.network.neutron [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.236268] env[62000]: DEBUG nova.network.neutron [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Updated VIF entry in instance network info cache for port 46a0ba05-811c-4e47-ad7c-0b6021236d10. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 691.236268] env[62000]: DEBUG nova.network.neutron [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Updating instance_info_cache with network_info: [{"id": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "address": "fa:16:3e:8e:e8:bf", "network": {"id": "c0a7cb0d-5e6f-4f14-81c5-ccab76f340e2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-151434811-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4300067ba40e4949861567c49e3f8093", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "796c7fcb-00fd-4692-a44b-7ec550201e86", "external-id": "nsx-vlan-transportzone-42", "segmentation_id": 42, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap46a0ba05-81", "ovs_interfaceid": "46a0ba05-811c-4e47-ad7c-0b6021236d10", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.374170] env[62000]: DEBUG nova.scheduler.client.report [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.398026] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d43ff8-8bf3-952b-b2c3-b1af52850904, 'name': SearchDatastore_Task, 'duration_secs': 0.009482} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.399036] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ac15158-6823-48f1-91cf-e6a34b5bd4dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.404672] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 691.404672] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527f5c49-d8f9-37b7-8fa6-2ee0a925fd35" [ 691.404672] env[62000]: _type = "Task" [ 691.404672] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.414150] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527f5c49-d8f9-37b7-8fa6-2ee0a925fd35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 691.514524] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 691.606055] env[62000]: INFO nova.compute.manager [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] [instance: 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4] Took 1.02 seconds to deallocate network for instance. [ 691.740228] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef44cf70-f7e5-4d07-a786-0850775ad0dd req-8c53aca7-de92-4616-8edc-474047c4284c service nova] Releasing lock "refresh_cache-1419e5a8-a78f-44f1-bfc4-00fcf176bf35" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.850029] env[62000]: DEBUG nova.network.neutron [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Updating instance_info_cache with network_info: [{"id": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "address": "fa:16:3e:38:58:44", "network": {"id": "6f0af99f-913d-4992-b6fb-1a20b020f2e7", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-342726683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076d68ba0f9d4cdbbfbdeb879a222c08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7bb956-76", "ovs_interfaceid": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.880251] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.880251] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.880251] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 27.731s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.925216] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527f5c49-d8f9-37b7-8fa6-2ee0a925fd35, 'name': SearchDatastore_Task, 'duration_secs': 0.00891} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 691.925445] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.925691] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 1419e5a8-a78f-44f1-bfc4-00fcf176bf35/1419e5a8-a78f-44f1-bfc4-00fcf176bf35.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 691.925946] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8864a8c4-0b32-4d39-8b09-82a116bd93d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.937256] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 691.937256] env[62000]: value = "task-881853" [ 691.937256] env[62000]: _type = "Task" [ 691.937256] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 691.945306] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.355318] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Releasing lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.355715] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Instance network_info: |[{"id": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "address": "fa:16:3e:38:58:44", "network": {"id": "6f0af99f-913d-4992-b6fb-1a20b020f2e7", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-342726683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076d68ba0f9d4cdbbfbdeb879a222c08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7bb956-76", "ovs_interfaceid": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 692.356228] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:58:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b4d548e7-d762-406a-bb2d-dc7168a8ca67', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ff7bb956-76bd-4472-99b0-ae8305b3f7b0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 692.364119] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Creating folder: Project (076d68ba0f9d4cdbbfbdeb879a222c08). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.364380] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be066278-e52b-4a21-9096-025a0b41e4fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.375036] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Created folder: Project (076d68ba0f9d4cdbbfbdeb879a222c08) in parent group-v201431. [ 692.375157] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Creating folder: Instances. Parent ref: group-v201444. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 692.375399] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86e2362a-2fcd-42e3-b4ad-8d7c0f959be3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.390305] env[62000]: DEBUG nova.compute.utils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.391508] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Created folder: Instances in parent group-v201444. [ 692.391590] env[62000]: DEBUG oslo.service.loopingcall [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 692.392118] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.392298] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 692.394791] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 692.395231] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9c3d9ec8-07d4-46bd-b54f-6e335a90f4b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.418443] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 692.418443] env[62000]: value = "task-881856" [ 692.418443] env[62000]: _type = "Task" [ 692.418443] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.426473] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881856, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.449859] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.429903} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.449859] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 1419e5a8-a78f-44f1-bfc4-00fcf176bf35/1419e5a8-a78f-44f1-bfc4-00fcf176bf35.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 692.449859] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 692.449859] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c65bc9a0-7403-47b7-b711-f30b46cee924 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.455079] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 692.455079] env[62000]: value = "task-881857" [ 692.455079] env[62000]: _type = "Task" [ 692.455079] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.465588] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.481537] env[62000]: DEBUG nova.policy [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8352c796b8764961a489857ca33b5317', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '99c8317210514ebaa8a975fe063b5378', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.630449] env[62000]: DEBUG nova.compute.manager [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Received event network-vif-plugged-ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 692.630449] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Acquiring lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.630449] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.630449] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.630449] env[62000]: DEBUG nova.compute.manager [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] No waiting events found dispatching network-vif-plugged-ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 692.630978] env[62000]: WARNING nova.compute.manager [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Received unexpected event network-vif-plugged-ff7bb956-76bd-4472-99b0-ae8305b3f7b0 for instance with vm_state building and task_state spawning. [ 692.630978] env[62000]: DEBUG nova.compute.manager [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Received event network-changed-ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 692.630978] env[62000]: DEBUG nova.compute.manager [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Refreshing instance network info cache due to event network-changed-ff7bb956-76bd-4472-99b0-ae8305b3f7b0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 692.630978] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Acquiring lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.630978] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Acquired lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.633240] env[62000]: DEBUG nova.network.neutron [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Refreshing network info cache for port ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 692.656921] env[62000]: INFO nova.scheduler.client.report [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Deleted allocations for instance 7ff67667-32ce-48ee-bae6-32a7b7d0bfd4 [ 692.785173] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d78dd4-2ac0-4cff-8bac-3eceec2a4ae2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.793263] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12bd80d3-17ca-4aec-87d2-81d8806c3c59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.826129] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e9635e-7c64-4b58-a724-6e322a87c9ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.833724] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d131201d-631c-4f8e-abb5-db84e8461664 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.847891] env[62000]: DEBUG nova.compute.provider_tree [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.897366] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.926013] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Successfully created port: 1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.931583] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881856, 'name': CreateVM_Task, 'duration_secs': 0.375595} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.934948] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 692.934948] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.934948] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.934948] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 692.934948] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4c567af-03b4-47cf-9df6-ed8bd9462c77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.940017] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 692.940017] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524b1fed-86a7-4372-6595-7bd779d844fe" [ 692.940017] env[62000]: _type = "Task" [ 692.940017] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.946852] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524b1fed-86a7-4372-6595-7bd779d844fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.963400] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109489} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.963677] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 692.964463] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c31538c-eb9f-427f-9f5c-cf0ac5372d15 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.987690] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Reconfiguring VM instance instance-0000001f to attach disk [datastore1] 1419e5a8-a78f-44f1-bfc4-00fcf176bf35/1419e5a8-a78f-44f1-bfc4-00fcf176bf35.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 692.987690] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1a0f865-f8e3-4f8b-b2be-8ffb842fa65e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.008269] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 693.008269] env[62000]: value = "task-881858" [ 693.008269] env[62000]: _type = "Task" [ 693.008269] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.016968] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881858, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.172514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-355e6f95-1e10-4811-bad7-fe43865ab3d6 tempest-ListImageFiltersTestJSON-890623880 tempest-ListImageFiltersTestJSON-890623880-project-member] Lock "7ff67667-32ce-48ee-bae6-32a7b7d0bfd4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.784s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.320498] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.320733] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.353473] env[62000]: DEBUG nova.scheduler.client.report [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.451016] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524b1fed-86a7-4372-6595-7bd779d844fe, 'name': SearchDatastore_Task, 'duration_secs': 0.00982} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.453530] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.453767] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 693.453992] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.454152] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.454330] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 693.454987] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8d0e8a8b-8000-4cca-999d-52277d38332c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.463686] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 693.463862] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 693.464563] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3e9d4a3-08e5-429c-a918-05b9fb009658 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.472325] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 693.472325] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5264f32c-b900-79f1-71d7-97c530faa067" [ 693.472325] env[62000]: _type = "Task" [ 693.472325] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.480526] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5264f32c-b900-79f1-71d7-97c530faa067, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.518325] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881858, 'name': ReconfigVM_Task, 'duration_secs': 0.269662} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.518713] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Reconfigured VM instance instance-0000001f to attach disk [datastore1] 1419e5a8-a78f-44f1-bfc4-00fcf176bf35/1419e5a8-a78f-44f1-bfc4-00fcf176bf35.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 693.519233] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ef7c483e-b956-4505-a2b7-ffa7d9c3da9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.526512] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 693.526512] env[62000]: value = "task-881859" [ 693.526512] env[62000]: _type = "Task" [ 693.526512] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.533860] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881859, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.536306] env[62000]: DEBUG nova.network.neutron [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Updated VIF entry in instance network info cache for port ff7bb956-76bd-4472-99b0-ae8305b3f7b0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 693.536622] env[62000]: DEBUG nova.network.neutron [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Updating instance_info_cache with network_info: [{"id": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "address": "fa:16:3e:38:58:44", "network": {"id": "6f0af99f-913d-4992-b6fb-1a20b020f2e7", "bridge": "br-int", "label": "tempest-ImagesOneServerTestJSON-342726683-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "076d68ba0f9d4cdbbfbdeb879a222c08", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b4d548e7-d762-406a-bb2d-dc7168a8ca67", "external-id": "nsx-vlan-transportzone-796", "segmentation_id": 796, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapff7bb956-76", "ovs_interfaceid": "ff7bb956-76bd-4472-99b0-ae8305b3f7b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.676496] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 693.858641] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.978s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.859571] env[62000]: ERROR nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Traceback (most recent call last): [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.driver.spawn(context, instance, image_meta, [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 540, in spawn [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] vm_ref = self.build_virtual_machine(instance, [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.859571] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] for vif in network_info: [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return self._sync_wrapper(fn, *args, **kwargs) [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.wait() [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self[:] = self._gt.wait() [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return self._exit_event.wait() [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] current.throw(*self._exc) [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.859969] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] result = function(*args, **kwargs) [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] return func(*args, **kwargs) [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise e [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] nwinfo = self.network_api.allocate_for_instance( [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] created_port_ids = self._update_ports_for_instance( [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] with excutils.save_and_reraise_exception(): [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] self.force_reraise() [ 693.860377] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise self.value [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] updated_port = self._update_port( [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] _ensure_no_port_binding_failure(port) [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] raise exception.PortBindingFailed(port_id=port['id']) [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] nova.exception.PortBindingFailed: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. [ 693.860719] env[62000]: ERROR nova.compute.manager [instance: f36c4c0a-6777-468f-8798-44093023b2d1] [ 693.860719] env[62000]: DEBUG nova.compute.utils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.861992] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.056s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.863332] env[62000]: INFO nova.compute.claims [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.866207] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Build of instance f36c4c0a-6777-468f-8798-44093023b2d1 was re-scheduled: Binding failed for port adfb24f7-ba06-4faf-bb95-ebfcbc767b3f, please check neutron logs for more information. {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 693.866686] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Unplugging VIFs for instance {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 693.866914] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.867075] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquired lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.867238] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.907172] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.940939] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.941219] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.941377] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.941557] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.941709] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.942044] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.942339] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.942519] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.942687] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.942848] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.943027] env[62000]: DEBUG nova.virt.hardware [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.943892] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942ca009-2bee-40f3-9367-960554bad1f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.952899] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6866159b-2685-4d68-86b7-82a470d9513a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.983261] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5264f32c-b900-79f1-71d7-97c530faa067, 'name': SearchDatastore_Task, 'duration_secs': 0.007891} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.984720] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f18fd7b-a146-4dc5-b038-b40cad17c9fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.991197] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 693.991197] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a7e5be-4ee6-f3ac-b131-67fd5488b1d7" [ 693.991197] env[62000]: _type = "Task" [ 693.991197] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.999874] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a7e5be-4ee6-f3ac-b131-67fd5488b1d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.037396] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881859, 'name': Rename_Task, 'duration_secs': 0.130225} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.037669] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 694.037908] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7cf90f4-65a2-4a74-8be6-7c1746325702 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.039987] env[62000]: DEBUG oslo_concurrency.lockutils [req-78fb8296-4848-4c25-b978-380326276220 req-449c3fdd-1126-44ae-bced-ee8454a3d96d service nova] Releasing lock "refresh_cache-1298037f-9cb7-4e2b-b70f-2ab24efe9b91" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.045385] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 694.045385] env[62000]: value = "task-881860" [ 694.045385] env[62000]: _type = "Task" [ 694.045385] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.054688] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881860, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.202603] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.390419] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 694.487733] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.500875] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a7e5be-4ee6-f3ac-b131-67fd5488b1d7, 'name': SearchDatastore_Task, 'duration_secs': 0.008708} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.501189] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.501488] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 1298037f-9cb7-4e2b-b70f-2ab24efe9b91/1298037f-9cb7-4e2b-b70f-2ab24efe9b91.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 694.501776] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c6e588a0-6aa8-4129-881f-a121ecc36302 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.509222] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 694.509222] env[62000]: value = "task-881861" [ 694.509222] env[62000]: _type = "Task" [ 694.509222] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.518748] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881861, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.559726] env[62000]: DEBUG oslo_vmware.api [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881860, 'name': PowerOnVM_Task, 'duration_secs': 0.415014} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.559726] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 694.559726] env[62000]: INFO nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Took 7.46 seconds to spawn the instance on the hypervisor. [ 694.559726] env[62000]: DEBUG nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 694.559726] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd5c3f07-34ab-42be-9ed0-e80a5fd27a0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.835592] env[62000]: DEBUG nova.compute.manager [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Received event network-vif-plugged-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 694.836965] env[62000]: DEBUG oslo_concurrency.lockutils [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] Acquiring lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.837732] env[62000]: DEBUG oslo_concurrency.lockutils [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] Lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.838042] env[62000]: DEBUG oslo_concurrency.lockutils [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] Lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 694.838544] env[62000]: DEBUG nova.compute.manager [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] No waiting events found dispatching network-vif-plugged-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 694.839431] env[62000]: WARNING nova.compute.manager [req-15c37e09-92cf-4a5d-8770-de5b381bee4e req-91474978-90be-4f89-9810-8d81ac6d1041 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Received unexpected event network-vif-plugged-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 for instance with vm_state building and task_state spawning. [ 694.887315] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Successfully updated port: 1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 694.992631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Releasing lock "refresh_cache-f36c4c0a-6777-468f-8798-44093023b2d1" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.992915] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62000) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 694.993186] env[62000]: DEBUG nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.993408] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 695.015198] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.026757] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881861, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493172} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.029688] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 1298037f-9cb7-4e2b-b70f-2ab24efe9b91/1298037f-9cb7-4e2b-b70f-2ab24efe9b91.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 695.030232] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 695.030927] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f7753ed8-265c-4603-8f36-dfdfb9fd9673 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.041087] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 695.041087] env[62000]: value = "task-881862" [ 695.041087] env[62000]: _type = "Task" [ 695.041087] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.050293] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881862, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.076903] env[62000]: INFO nova.compute.manager [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Took 40.25 seconds to build instance. [ 695.248614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b82bf0fb-bb31-4f79-bae6-a9c8f2e613c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.256178] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0adaf0c-92a6-48d9-86e8-2eefa0dd4114 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.285092] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9973206-9a17-4cb9-86a0-f4d85d5454f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.292459] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffc73950-0b82-4715-aa2d-bb48d3fe264c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.305506] env[62000]: DEBUG nova.compute.provider_tree [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 695.389668] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.389830] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquired lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.389987] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.525202] env[62000]: DEBUG nova.network.neutron [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.551112] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881862, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072478} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 695.551395] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 695.553298] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0c9fc4-f398-45ea-9b07-fd4ca829de90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.575735] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Reconfiguring VM instance instance-00000020 to attach disk [datastore1] 1298037f-9cb7-4e2b-b70f-2ab24efe9b91/1298037f-9cb7-4e2b-b70f-2ab24efe9b91.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 695.576020] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c0f24ab-79d8-41c1-bc07-fe0a028850e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.590417] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b818055-d333-452c-be7f-b68fddefa7a3 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.852s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.597060] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 695.597060] env[62000]: value = "task-881863" [ 695.597060] env[62000]: _type = "Task" [ 695.597060] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 695.606667] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881863, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 695.811019] env[62000]: DEBUG nova.scheduler.client.report [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.944682] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.028395] env[62000]: INFO nova.compute.manager [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: f36c4c0a-6777-468f-8798-44093023b2d1] Took 1.03 seconds to deallocate network for instance. [ 696.093168] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 696.107612] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881863, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.119823] env[62000]: DEBUG nova.network.neutron [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Updating instance_info_cache with network_info: [{"id": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "address": "fa:16:3e:1d:b0:2e", "network": {"id": "cac9d05c-a922-4eb8-bbc0-7c71c9bc97f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1076136654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99c8317210514ebaa8a975fe063b5378", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69d412f5-01a9-4fed-8488-7b767a13a653", "external-id": "nsx-vlan-transportzone-444", "segmentation_id": 444, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4c2dd6-fe", "ovs_interfaceid": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.312945] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 696.313552] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 696.316433] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.474s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.317812] env[62000]: INFO nova.compute.claims [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.610891] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881863, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.619195] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.622736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Releasing lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.623041] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Instance network_info: |[{"id": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "address": "fa:16:3e:1d:b0:2e", "network": {"id": "cac9d05c-a922-4eb8-bbc0-7c71c9bc97f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1076136654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99c8317210514ebaa8a975fe063b5378", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69d412f5-01a9-4fed-8488-7b767a13a653", "external-id": "nsx-vlan-transportzone-444", "segmentation_id": 444, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4c2dd6-fe", "ovs_interfaceid": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 696.623425] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:b0:2e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '69d412f5-01a9-4fed-8488-7b767a13a653', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 696.630684] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Creating folder: Project (99c8317210514ebaa8a975fe063b5378). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 696.630924] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b4ba494-eb44-4cc5-a2d5-02c1637122c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.640356] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Created folder: Project (99c8317210514ebaa8a975fe063b5378) in parent group-v201431. [ 696.640529] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Creating folder: Instances. Parent ref: group-v201447. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 696.640731] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a7e4d215-2958-496f-ba1b-bc0245af28f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.649408] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Created folder: Instances in parent group-v201447. [ 696.649617] env[62000]: DEBUG oslo.service.loopingcall [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 696.649787] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 696.649970] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7f5feb9a-9a97-4bb4-9862-7d2388966c7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.667051] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 696.667051] env[62000]: value = "task-881866" [ 696.667051] env[62000]: _type = "Task" [ 696.667051] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 696.673832] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881866, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 696.822019] env[62000]: DEBUG nova.compute.utils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.825235] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.825443] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 696.886026] env[62000]: DEBUG nova.policy [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afdefedc02bc4b3e9b92336f57a9997b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e45b26d92ce540cea0f7cd27f52da2ec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.895878] env[62000]: DEBUG nova.compute.manager [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Received event network-changed-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 696.896184] env[62000]: DEBUG nova.compute.manager [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Refreshing instance network info cache due to event network-changed-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 696.896555] env[62000]: DEBUG oslo_concurrency.lockutils [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] Acquiring lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 696.896555] env[62000]: DEBUG oslo_concurrency.lockutils [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] Acquired lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 696.896817] env[62000]: DEBUG nova.network.neutron [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Refreshing network info cache for port 1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 697.065661] env[62000]: INFO nova.scheduler.client.report [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Deleted allocations for instance f36c4c0a-6777-468f-8798-44093023b2d1 [ 697.110671] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881863, 'name': ReconfigVM_Task, 'duration_secs': 1.036404} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.111051] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Reconfigured VM instance instance-00000020 to attach disk [datastore1] 1298037f-9cb7-4e2b-b70f-2ab24efe9b91/1298037f-9cb7-4e2b-b70f-2ab24efe9b91.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 697.111686] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ad479db4-40d1-4c9d-abd7-f4332a858d44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.119710] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 697.119710] env[62000]: value = "task-881867" [ 697.119710] env[62000]: _type = "Task" [ 697.119710] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.128647] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881867, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.181041] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881866, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.295238] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully created port: 3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.332023] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.578750] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b2c126c-f032-4cfd-9d37-6374fd32d5c2 tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "f36c4c0a-6777-468f-8798-44093023b2d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.137s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.629999] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881867, 'name': Rename_Task, 'duration_secs': 0.149564} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 697.630317] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 697.630552] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8cbf9d95-86cb-4070-afd6-e5019195a6f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.641509] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 697.641509] env[62000]: value = "task-881868" [ 697.641509] env[62000]: _type = "Task" [ 697.641509] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.645030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.645260] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.645449] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.645624] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.645785] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.651075] env[62000]: INFO nova.compute.manager [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Terminating instance [ 697.661595] env[62000]: DEBUG nova.compute.manager [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 697.662262] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 697.663284] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881868, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.663657] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ffe970b-1778-4fe1-b0e1-d1f93be9f2c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.670847] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 697.674015] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5e7e78f7-12c3-4475-8c6a-c1a7c1ebfc19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.682938] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881866, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.685078] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 697.685078] env[62000]: value = "task-881869" [ 697.685078] env[62000]: _type = "Task" [ 697.685078] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 697.693109] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 697.694653] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b84948-54cc-47ca-a337-92ff9033293b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.701187] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eab8d285-c185-476b-89b7-784da273d445 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.734764] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f831df-1820-4709-a124-0b29ee8ed420 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.742731] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55413db3-a81e-4a2f-9dbd-aaa2f0d5f5a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.755925] env[62000]: DEBUG nova.compute.provider_tree [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.759313] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully created port: 556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 697.806331] env[62000]: DEBUG nova.network.neutron [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Updated VIF entry in instance network info cache for port 1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 697.806331] env[62000]: DEBUG nova.network.neutron [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Updating instance_info_cache with network_info: [{"id": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "address": "fa:16:3e:1d:b0:2e", "network": {"id": "cac9d05c-a922-4eb8-bbc0-7c71c9bc97f7", "bridge": "br-int", "label": "tempest-InstanceActionsV221TestJSON-1076136654-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "99c8317210514ebaa8a975fe063b5378", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "69d412f5-01a9-4fed-8488-7b767a13a653", "external-id": "nsx-vlan-transportzone-444", "segmentation_id": 444, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f4c2dd6-fe", "ovs_interfaceid": "1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.081782] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 698.088212] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully created port: 10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 698.157185] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881868, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.183713] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881866, 'name': CreateVM_Task, 'duration_secs': 1.307115} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.183713] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 698.184050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.184633] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.184633] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 698.184823] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b63ed54-b380-47dd-8a21-381e0d1ed454 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.194491] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 698.194491] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526952b8-be3b-790f-9ba9-a018009aaa24" [ 698.194491] env[62000]: _type = "Task" [ 698.194491] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.197999] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881869, 'name': PowerOffVM_Task, 'duration_secs': 0.208102} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.201075] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 698.201264] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 698.201819] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3abd7714-30de-4399-975a-7ccf94576416 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.208327] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526952b8-be3b-790f-9ba9-a018009aaa24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.259148] env[62000]: DEBUG nova.scheduler.client.report [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 698.272107] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 698.272356] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 698.277020] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Deleting the datastore file [datastore1] 1419e5a8-a78f-44f1-bfc4-00fcf176bf35 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 698.277020] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07404675-63db-4cd0-b2d3-1452dba741e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.279679] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for the task: (returnval){ [ 698.279679] env[62000]: value = "task-881871" [ 698.279679] env[62000]: _type = "Task" [ 698.279679] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.290630] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.308961] env[62000]: DEBUG oslo_concurrency.lockutils [req-9bc651cb-58b5-49da-bbf1-6b91ee11f049 req-f83a5952-051d-4167-9445-ebc772cfe6f9 service nova] Releasing lock "refresh_cache-762e429f-f690-43c4-95eb-877caf1cdad7" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.343151] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 698.370052] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 698.370327] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 698.370486] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 698.370662] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 698.370806] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 698.370950] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 698.371422] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 698.371732] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 698.372044] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 698.372553] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 698.372845] env[62000]: DEBUG nova.virt.hardware [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 698.374292] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7fa4f4-a6a0-4150-a616-56c08508b331 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.383213] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aadab28f-92fd-4b16-ab1f-98bc8936ef6f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.604859] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.652214] env[62000]: DEBUG oslo_vmware.api [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881868, 'name': PowerOnVM_Task, 'duration_secs': 0.56314} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.652482] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 698.652675] env[62000]: INFO nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 9.11 seconds to spawn the instance on the hypervisor. [ 698.652855] env[62000]: DEBUG nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 698.653642] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff826c8-dc53-48fe-9635-1f694c0e2049 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.707774] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526952b8-be3b-790f-9ba9-a018009aaa24, 'name': SearchDatastore_Task, 'duration_secs': 0.013299} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.708077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.708306] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 698.708530] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.708672] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.708846] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 698.709110] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8cefb8dc-5227-4dfa-b460-bd74166959e1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.717693] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 698.717873] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 698.718580] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a921d11-fc28-4ffb-a97b-9bd7d00a5a2d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.723852] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 698.723852] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959179-4ea6-f5c7-8b9a-79ee5990f11d" [ 698.723852] env[62000]: _type = "Task" [ 698.723852] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 698.731345] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959179-4ea6-f5c7-8b9a-79ee5990f11d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 698.768607] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.768607] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.771076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.073s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.772948] env[62000]: INFO nova.compute.claims [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.789329] env[62000]: DEBUG oslo_vmware.api [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Task: {'id': task-881871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251373} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 698.789635] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 698.789851] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 698.790090] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 698.790307] env[62000]: INFO nova.compute.manager [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Took 1.13 seconds to destroy the instance on the hypervisor. [ 698.790581] env[62000]: DEBUG oslo.service.loopingcall [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 698.790800] env[62000]: DEBUG nova.compute.manager [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 698.790932] env[62000]: DEBUG nova.network.neutron [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 699.088423] env[62000]: DEBUG nova.compute.manager [req-791d1a17-581f-4fcf-8378-3a52e4dd5e14 req-76ea6270-db2a-400e-9e72-0c306194ac30 service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Received event network-vif-deleted-46a0ba05-811c-4e47-ad7c-0b6021236d10 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 699.088423] env[62000]: INFO nova.compute.manager [req-791d1a17-581f-4fcf-8378-3a52e4dd5e14 req-76ea6270-db2a-400e-9e72-0c306194ac30 service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Neutron deleted interface 46a0ba05-811c-4e47-ad7c-0b6021236d10; detaching it from the instance and deleting it from the info cache [ 699.088423] env[62000]: DEBUG nova.network.neutron [req-791d1a17-581f-4fcf-8378-3a52e4dd5e14 req-76ea6270-db2a-400e-9e72-0c306194ac30 service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.169813] env[62000]: INFO nova.compute.manager [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 42.43 seconds to build instance. [ 699.235018] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959179-4ea6-f5c7-8b9a-79ee5990f11d, 'name': SearchDatastore_Task, 'duration_secs': 0.010108} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.235018] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b23ad9f-3de6-42b0-9a78-78823d88d717 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.240984] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 699.240984] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5247cc53-9c8c-00d2-d6f8-ae1bef36e0aa" [ 699.240984] env[62000]: _type = "Task" [ 699.240984] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.251439] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5247cc53-9c8c-00d2-d6f8-ae1bef36e0aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.277320] env[62000]: DEBUG nova.compute.utils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 699.281433] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 699.519231] env[62000]: DEBUG nova.network.neutron [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.590750] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a392eb7-c6d5-41ec-9563-f26d26b75ce9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.605345] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77d92d4-7de7-4a4c-b39e-9db3f3a447d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.627661] env[62000]: DEBUG nova.compute.manager [req-791d1a17-581f-4fcf-8378-3a52e4dd5e14 req-76ea6270-db2a-400e-9e72-0c306194ac30 service nova] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Detach interface failed, port_id=46a0ba05-811c-4e47-ad7c-0b6021236d10, reason: Instance 1419e5a8-a78f-44f1-bfc4-00fcf176bf35 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 699.672101] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f309c94a-f615-4de7-a31a-a7661eadf7c2 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.318s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.752625] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5247cc53-9c8c-00d2-d6f8-ae1bef36e0aa, 'name': SearchDatastore_Task, 'duration_secs': 0.018269} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 699.752930] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.753221] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 762e429f-f690-43c4-95eb-877caf1cdad7/762e429f-f690-43c4-95eb-877caf1cdad7.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 699.753488] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bd65789d-3524-422f-ba1d-196901ba7844 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.760390] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 699.760390] env[62000]: value = "task-881872" [ 699.760390] env[62000]: _type = "Task" [ 699.760390] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.768050] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881872, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.783363] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.870386] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully updated port: 3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 700.021092] env[62000]: INFO nova.compute.manager [-] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Took 1.23 seconds to deallocate network for instance. [ 700.132479] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0db960-2a38-46ac-b0e4-d49eb8c1b2b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.141288] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db774079-8248-42a8-9bc5-a07e71fbe582 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.173949] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19a413d-c04e-43c4-9423-e1503d02f91f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.176948] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 700.185637] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-575f2812-dc29-431e-be6c-3d4595553437 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.201923] env[62000]: DEBUG nova.compute.provider_tree [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 700.269743] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881872, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446851} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.270059] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 762e429f-f690-43c4-95eb-877caf1cdad7/762e429f-f690-43c4-95eb-877caf1cdad7.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 700.270821] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 700.270821] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b304994e-96d9-4172-b89a-d4648a2a2a6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.278095] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 700.278095] env[62000]: value = "task-881873" [ 700.278095] env[62000]: _type = "Task" [ 700.278095] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.285839] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881873, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.530256] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.696885] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.704926] env[62000]: DEBUG nova.scheduler.client.report [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.788687] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881873, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062264} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 700.788962] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 700.789757] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b015a1-477c-466f-a064-6b44bc03900e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.805861] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 700.815953] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Reconfiguring VM instance instance-00000021 to attach disk [datastore1] 762e429f-f690-43c4-95eb-877caf1cdad7/762e429f-f690-43c4-95eb-877caf1cdad7.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 700.817865] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2414ebfc-8424-4899-9601-643220d56313 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.839557] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 700.839830] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 700.840007] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 700.840201] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 700.840345] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 700.840489] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 700.840689] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 700.840844] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 700.841015] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 700.841183] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 700.841350] env[62000]: DEBUG nova.virt.hardware [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 700.842989] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2214c92-36c0-4eec-bdf9-d0df1d495f00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.847497] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 700.847497] env[62000]: value = "task-881874" [ 700.847497] env[62000]: _type = "Task" [ 700.847497] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.854786] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5252aed8-1680-4277-8e7c-45f082a3ec02 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.861675] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881874, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 700.871705] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 700.877588] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Creating folder: Project (c700dfde5a7047778a5ac7af481f00f8). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.877940] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb7a4851-a067-417b-880c-26b24417298d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.888280] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Created folder: Project (c700dfde5a7047778a5ac7af481f00f8) in parent group-v201431. [ 700.889268] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Creating folder: Instances. Parent ref: group-v201450. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 700.889268] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0b94a63-821e-427c-936b-736586664918 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.898007] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Created folder: Instances in parent group-v201450. [ 700.898317] env[62000]: DEBUG oslo.service.loopingcall [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.898545] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 700.898779] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5c9eece8-a87f-42ba-a0a8-83d075b83985 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.915648] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 700.915648] env[62000]: value = "task-881877" [ 700.915648] env[62000]: _type = "Task" [ 700.915648] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 700.923729] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881877, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.125931] env[62000]: DEBUG nova.compute.manager [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-plugged-3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.126219] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.126438] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.126607] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.126776] env[62000]: DEBUG nova.compute.manager [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] No waiting events found dispatching network-vif-plugged-3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 701.126938] env[62000]: WARNING nova.compute.manager [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received unexpected event network-vif-plugged-3d6933ed-ca06-4521-a288-8c4c4cacefd1 for instance with vm_state building and task_state spawning. [ 701.127109] env[62000]: DEBUG nova.compute.manager [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-changed-3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 701.127265] env[62000]: DEBUG nova.compute.manager [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing instance network info cache due to event network-changed-3d6933ed-ca06-4521-a288-8c4c4cacefd1. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 701.127444] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Acquiring lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.127576] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Acquired lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.127728] env[62000]: DEBUG nova.network.neutron [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing network info cache for port 3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 701.209818] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.439s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.210390] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 701.213182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.673s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.214790] env[62000]: INFO nova.compute.claims [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.249127] env[62000]: DEBUG nova.compute.manager [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 701.250271] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7de1159-3126-4cdc-960a-86b58bbc38f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.358197] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881874, 'name': ReconfigVM_Task, 'duration_secs': 0.290047} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.358499] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Reconfigured VM instance instance-00000021 to attach disk [datastore1] 762e429f-f690-43c4-95eb-877caf1cdad7/762e429f-f690-43c4-95eb-877caf1cdad7.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 701.359138] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-485fd004-a4ff-456d-bb42-8c6adfb5c505 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.365593] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 701.365593] env[62000]: value = "task-881878" [ 701.365593] env[62000]: _type = "Task" [ 701.365593] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.373953] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881878, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.428136] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881877, 'name': CreateVM_Task, 'duration_secs': 0.486127} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.428347] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 701.429244] env[62000]: DEBUG oslo_vmware.service [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92c07292-75bd-43da-996a-e5edbd67c62f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.434813] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.434989] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.435379] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 701.435618] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c188de2-1ebf-48b7-812d-54b004ece2a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.440140] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 701.440140] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5262ba3e-dd07-e1f8-1e40-bffa16c54dd0" [ 701.440140] env[62000]: _type = "Task" [ 701.440140] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.447267] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5262ba3e-dd07-e1f8-1e40-bffa16c54dd0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.663769] env[62000]: DEBUG nova.network.neutron [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.719228] env[62000]: DEBUG nova.compute.utils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.724024] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.724024] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 701.762655] env[62000]: INFO nova.compute.manager [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] instance snapshotting [ 701.765424] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5e18bd-97a0-4922-9281-d5502c7143dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.784954] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fc071af-a4f4-4150-9ac9-0b8f41b9a482 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.819606] env[62000]: DEBUG nova.policy [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '041d24bce7f14a24b84fa108f731283b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c93358b7815d4b5ea3b188965c276329', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.845823] env[62000]: DEBUG nova.network.neutron [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.875518] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881878, 'name': Rename_Task, 'duration_secs': 0.13709} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 701.875608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 701.875848] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f5ee3446-1a82-49ef-a285-facd601a4fd9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.884181] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 701.884181] env[62000]: value = "task-881879" [ 701.884181] env[62000]: _type = "Task" [ 701.884181] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.894028] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881879, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 701.951985] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.952280] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 701.952517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 701.952687] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 701.952875] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 701.953147] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebbb245e-8432-48a4-8ee7-26df3f8e095b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.971505] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 701.971701] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 701.972828] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c9332d7-ec8d-45e5-add3-a2f0dc57c1c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.979842] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4d34d2a4-33ab-483d-9a31-03930d100518 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.985456] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 701.985456] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523dfe2a-01e8-1b65-c7b1-9b6f1af2301c" [ 701.985456] env[62000]: _type = "Task" [ 701.985456] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 701.996219] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523dfe2a-01e8-1b65-c7b1-9b6f1af2301c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.219702] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully updated port: 556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 702.228364] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.297918] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 702.297918] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-800d62bd-2da5-4dce-a0f9-5a20b7b2d625 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.305525] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 702.305525] env[62000]: value = "task-881880" [ 702.305525] env[62000]: _type = "Task" [ 702.305525] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 702.317657] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881880, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.342850] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Successfully created port: d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 702.348747] env[62000]: DEBUG oslo_concurrency.lockutils [req-45ffde64-45a9-4d81-bd32-8bca6ee35098 req-cf5e721f-9d02-42b4-8b0c-98154833f3aa service nova] Releasing lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 702.393072] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881879, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 702.500075] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 702.500350] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Creating directory with path [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 702.501565] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53622ae9-b8b2-4808-9bd0-ef3932d5d71e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.515633] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Created directory with path [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 702.515843] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Fetch image to [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 702.516092] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Downloading image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk on the data store datastore2 {{(pid=62000) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 702.516796] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c05189d-73e7-4f6e-83c4-4c0187ad11a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.523785] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fee401a-a6a5-4680-8c24-658630dfd9df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.536190] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a47a2d2-a385-4523-8267-a3bd848cef5b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.573881] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edffc4b-ba70-4904-b5c3-83f14618891b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.582335] env[62000]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-3cef9b7c-6e51-40ac-91d4-d1b8456f8420 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.591193] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad26febc-d5af-4a7b-ad62-2957dc7b5078 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.598769] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0234938-4c33-41e0-a53d-eda38807b4cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.631771] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d076fa3-eade-483c-9ae6-e6249d620e4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.636076] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Downloading image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to the data store datastore2 {{(pid=62000) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 702.643015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1110cd25-9543-4822-8f79-38cb75529671 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.656414] env[62000]: DEBUG nova.compute.provider_tree [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 702.696033] env[62000]: DEBUG oslo_vmware.rw_handles [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 702.819366] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881880, 'name': CreateSnapshot_Task, 'duration_secs': 0.465848} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.820266] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 702.820712] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-124c5ec4-0241-4e17-b4a4-e75cbbcbcda3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.896988] env[62000]: DEBUG oslo_vmware.api [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881879, 'name': PowerOnVM_Task, 'duration_secs': 0.73684} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 702.897198] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 702.897346] env[62000]: INFO nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Took 8.99 seconds to spawn the instance on the hypervisor. [ 702.897524] env[62000]: DEBUG nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 702.898548] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e0afd8-4553-4066-b9b3-fe0ca98417c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.156906] env[62000]: DEBUG nova.compute.manager [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-plugged-556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.157149] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.157358] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.157524] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.157686] env[62000]: DEBUG nova.compute.manager [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] No waiting events found dispatching network-vif-plugged-556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 703.157863] env[62000]: WARNING nova.compute.manager [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received unexpected event network-vif-plugged-556236e8-eb70-4dc1-8fbd-5090a7df2605 for instance with vm_state building and task_state spawning. [ 703.158036] env[62000]: DEBUG nova.compute.manager [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-changed-556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 703.158193] env[62000]: DEBUG nova.compute.manager [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing instance network info cache due to event network-changed-556236e8-eb70-4dc1-8fbd-5090a7df2605. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 703.158370] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Acquiring lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.158503] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Acquired lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.158768] env[62000]: DEBUG nova.network.neutron [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing network info cache for port 556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 703.163253] env[62000]: DEBUG nova.scheduler.client.report [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 703.260659] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 703.282912] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 703.283181] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 703.283358] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 703.283555] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 703.283705] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 703.283849] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 703.284065] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 703.284229] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 703.284397] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 703.284558] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 703.284728] env[62000]: DEBUG nova.virt.hardware [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 703.285661] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-355c9544-db80-4d09-a30d-5a5ada85749b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.300148] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a59e103-c8a8-437e-b7b4-e2c9ad48c57a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.340663] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 703.340885] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fb0c3316-8556-478c-a9c3-100b23016931 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.348989] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 703.348989] env[62000]: value = "task-881881" [ 703.348989] env[62000]: _type = "Task" [ 703.348989] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.357025] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881881, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.417304] env[62000]: INFO nova.compute.manager [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Took 42.19 seconds to build instance. [ 703.559556] env[62000]: DEBUG oslo_vmware.rw_handles [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 703.559556] env[62000]: DEBUG oslo_vmware.rw_handles [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 703.670711] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.457s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.671380] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 703.675429] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.271s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.677387] env[62000]: INFO nova.compute.claims [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 703.685682] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Downloaded image file data 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk on the data store datastore2 {{(pid=62000) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 703.687906] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 703.688172] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Copying Virtual Disk [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk to [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 703.688439] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1cec76f1-882c-46e5-97cf-eff4a1bb0461 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.695943] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 703.695943] env[62000]: value = "task-881882" [ 703.695943] env[62000]: _type = "Task" [ 703.695943] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 703.705225] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881882, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.739537] env[62000]: DEBUG nova.network.neutron [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 703.859197] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881881, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 703.922990] env[62000]: DEBUG nova.network.neutron [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.924086] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb54886b-b8e0-4574-a773-84bcb7d7cbbc tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.066s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.156513] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Successfully updated port: d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.181478] env[62000]: DEBUG nova.compute.utils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 704.187031] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 704.187031] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 704.206585] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881882, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.299028] env[62000]: DEBUG nova.policy [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c0bb2b11b41d4e54870be3b6fdb01a28', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76de44563cdd4a3883d5153555c2e48e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 704.362320] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881881, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.425849] env[62000]: DEBUG oslo_concurrency.lockutils [req-a86cf5a8-1ffc-400b-a37c-95700b38f4e2 req-8f4fc7cf-002c-4f23-a0d6-6222dc8d9582 service nova] Releasing lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.428605] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 704.549521] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "762e429f-f690-43c4-95eb-877caf1cdad7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.549814] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.550031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.551315] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.551315] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 704.553443] env[62000]: INFO nova.compute.manager [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Terminating instance [ 704.555651] env[62000]: DEBUG nova.compute.manager [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.555839] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 704.556667] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167e422f-7f9e-4bae-b3dc-eefe1bca77fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.564669] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 704.564669] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6fab7c80-f3da-484d-a0d2-ac58d71f5800 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.570899] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 704.570899] env[62000]: value = "task-881883" [ 704.570899] env[62000]: _type = "Task" [ 704.570899] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.578629] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881883, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.666856] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.666856] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.666856] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 704.668858] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Successfully updated port: 10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 704.696467] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 704.709333] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881882, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.928134} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.709810] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Copied Virtual Disk [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk to [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 704.710208] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleting the datastore file [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/tmp-sparse.vmdk {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 704.710482] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c7f538fb-4eef-4b12-be86-be395ec5b631 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.718236] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 704.718236] env[62000]: value = "task-881884" [ 704.718236] env[62000]: _type = "Task" [ 704.718236] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 704.728014] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881884, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 704.856976] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Successfully created port: c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 704.865944] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881881, 'name': CloneVM_Task, 'duration_secs': 1.425558} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 704.866219] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Created linked-clone VM from snapshot [ 704.866971] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b737c7b5-4635-4959-a1ab-6777610c104f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.874578] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Uploading image 3f9a0b46-aee3-432d-9f8a-fbc1a16bd49e {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 704.911624] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 704.911624] env[62000]: value = "vm-201454" [ 704.911624] env[62000]: _type = "VirtualMachine" [ 704.911624] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 704.912322] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-9f83209a-d9ad-43b7-a58c-7ac51c3b4037 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.921609] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lease: (returnval){ [ 704.921609] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523947dc-a253-31cd-be0b-a7bf955df40d" [ 704.921609] env[62000]: _type = "HttpNfcLease" [ 704.921609] env[62000]: } obtained for exporting VM: (result){ [ 704.921609] env[62000]: value = "vm-201454" [ 704.921609] env[62000]: _type = "VirtualMachine" [ 704.921609] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 704.921856] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the lease: (returnval){ [ 704.921856] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523947dc-a253-31cd-be0b-a7bf955df40d" [ 704.921856] env[62000]: _type = "HttpNfcLease" [ 704.921856] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 704.929673] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 704.929673] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523947dc-a253-31cd-be0b-a7bf955df40d" [ 704.929673] env[62000]: _type = "HttpNfcLease" [ 704.929673] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 704.946329] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.060486] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fbb00a-78ad-4898-b18d-acda29b867b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.068470] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-550fb802-6ebf-4c75-947b-422fc3b0c670 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.079086] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881883, 'name': PowerOffVM_Task, 'duration_secs': 0.197648} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.104863] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 705.105089] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 705.105501] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9fedc5c3-01d0-4d98-911e-07ed14e78523 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.107457] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53082cd9-0992-455d-affb-7bfe096636bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.115618] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ded62a3-59b9-4914-aaa5-b18ae4ff5822 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.128723] env[62000]: DEBUG nova.compute.provider_tree [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 705.172349] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.172613] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.172937] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 705.186027] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 705.186280] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 705.186380] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Deleting the datastore file [datastore1] 762e429f-f690-43c4-95eb-877caf1cdad7 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.186610] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-899fd1c0-a867-428c-af5e-4f67ac153b9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.193774] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for the task: (returnval){ [ 705.193774] env[62000]: value = "task-881887" [ 705.193774] env[62000]: _type = "Task" [ 705.193774] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.205112] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881887, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.207730] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Received event network-vif-plugged-d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.207935] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.208149] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.208315] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.208481] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] No waiting events found dispatching network-vif-plugged-d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 705.208643] env[62000]: WARNING nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Received unexpected event network-vif-plugged-d725993d-bf92-4c34-9f24-1c0ed6557610 for instance with vm_state building and task_state spawning. [ 705.208803] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Received event network-changed-d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 705.208952] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Refreshing instance network info cache due to event network-changed-d725993d-bf92-4c34-9f24-1c0ed6557610. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 705.209131] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquiring lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 705.209888] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.226772] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881884, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.105985} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.227288] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.227507] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Moving file from [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 to [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72. {{(pid=62000) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 705.227767] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-c2be2c63-667f-4656-8137-468851d4ea68 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.236127] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 705.236127] env[62000]: value = "task-881888" [ 705.236127] env[62000]: _type = "Task" [ 705.236127] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.244412] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881888, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.347074] env[62000]: DEBUG nova.network.neutron [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating instance_info_cache with network_info: [{"id": "d725993d-bf92-4c34-9f24-1c0ed6557610", "address": "fa:16:3e:d1:f7:5d", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd725993d-bf", "ovs_interfaceid": "d725993d-bf92-4c34-9f24-1c0ed6557610", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.429732] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 705.429732] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523947dc-a253-31cd-be0b-a7bf955df40d" [ 705.429732] env[62000]: _type = "HttpNfcLease" [ 705.429732] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 705.430077] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 705.430077] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523947dc-a253-31cd-be0b-a7bf955df40d" [ 705.430077] env[62000]: _type = "HttpNfcLease" [ 705.430077] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 705.430821] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3faf9fe6-71b7-4f10-94b2-1ff503656263 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.438717] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 705.438904] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 705.535844] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-c3d59865-aa5f-4be2-89fd-5f58223299e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.631582] env[62000]: DEBUG nova.scheduler.client.report [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.705291] env[62000]: DEBUG oslo_vmware.api [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Task: {'id': task-881887, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.395301} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.705537] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 705.705717] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 705.705890] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 705.706074] env[62000]: INFO nova.compute.manager [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Took 1.15 seconds to destroy the instance on the hypervisor. [ 705.706351] env[62000]: DEBUG oslo.service.loopingcall [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.706561] env[62000]: DEBUG nova.compute.manager [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 705.706657] env[62000]: DEBUG nova.network.neutron [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 705.709097] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 705.732645] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 705.732645] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 705.732645] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 705.732854] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 705.732854] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 705.732854] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 705.732967] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 705.733358] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 705.733532] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 705.733695] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 705.733927] env[62000]: DEBUG nova.virt.hardware [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 705.735102] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a45fd34f-09fd-4daf-a172-6c4e03115d17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.756186] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1041f2a2-08e0-43a1-8eb6-8c10bd32b55f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.760967] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881888, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.044127} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 705.761979] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.763761] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] File moved {{(pid=62000) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 705.764113] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Cleaning up location [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 705.764358] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleting the datastore file [datastore2] vmware_temp/2c7674f2-c09a-403b-ba40-74fc0ae575db {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 705.765320] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-97e9448e-ebd8-4d75-bc40-075dc26da075 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.780469] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 705.780469] env[62000]: value = "task-881889" [ 705.780469] env[62000]: _type = "Task" [ 705.780469] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.788437] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881889, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 705.849492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.849854] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance network_info: |[{"id": "d725993d-bf92-4c34-9f24-1c0ed6557610", "address": "fa:16:3e:d1:f7:5d", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd725993d-bf", "ovs_interfaceid": "d725993d-bf92-4c34-9f24-1c0ed6557610", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 705.850371] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquired lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 705.850587] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Refreshing network info cache for port d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 705.851920] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:f7:5d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '360308f4-9d0a-4ec2-8bcf-44891f452847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd725993d-bf92-4c34-9f24-1c0ed6557610', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 705.859316] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Creating folder: Project (c93358b7815d4b5ea3b188965c276329). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.860227] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b6357dd9-7376-4414-b645-4df0fd36ce25 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.874367] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Created folder: Project (c93358b7815d4b5ea3b188965c276329) in parent group-v201431. [ 705.874567] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Creating folder: Instances. Parent ref: group-v201455. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 705.874816] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-db944cce-0922-4b52-b52b-0c23075bbbb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.888294] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Created folder: Instances in parent group-v201455. [ 705.888294] env[62000]: DEBUG oslo.service.loopingcall [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 705.888294] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 705.888294] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-05c88da7-af52-4279-b993-e2a71117e9bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 705.906571] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 705.906571] env[62000]: value = "task-881892" [ 705.906571] env[62000]: _type = "Task" [ 705.906571] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 705.914950] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881892, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.139072] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.139751] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 706.146210] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.944s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 706.147912] env[62000]: INFO nova.compute.claims [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.292496] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881889, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.02768} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.292773] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 706.293611] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5877f1f4-9d32-4211-8415-306376e0506c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.299853] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 706.299853] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52568d15-e6ef-a19e-5ba0-a1830de377f9" [ 706.299853] env[62000]: _type = "Task" [ 706.299853] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.308534] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52568d15-e6ef-a19e-5ba0-a1830de377f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.417448] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881892, 'name': CreateVM_Task, 'duration_secs': 0.320683} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.417716] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 706.418282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.418447] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.418766] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 706.419014] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c321adfc-d9b4-4f6b-8432-4130d6850c86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.423591] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 706.423591] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e662e4-8a29-3f6f-89b8-ca8ead986316" [ 706.423591] env[62000]: _type = "Task" [ 706.423591] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.432788] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e662e4-8a29-3f6f-89b8-ca8ead986316, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.655075] env[62000]: DEBUG nova.compute.utils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 706.658248] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 706.658422] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 706.675985] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updated VIF entry in instance network info cache for port d725993d-bf92-4c34-9f24-1c0ed6557610. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 706.676535] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating instance_info_cache with network_info: [{"id": "d725993d-bf92-4c34-9f24-1c0ed6557610", "address": "fa:16:3e:d1:f7:5d", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd725993d-bf", "ovs_interfaceid": "d725993d-bf92-4c34-9f24-1c0ed6557610", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.814292] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52568d15-e6ef-a19e-5ba0-a1830de377f9, 'name': SearchDatastore_Task, 'duration_secs': 0.015509} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.814586] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.814848] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 5e80a5c8-030c-4ad8-90c4-26136fa39d71/5e80a5c8-030c-4ad8-90c4-26136fa39d71.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 706.818543] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-367c1c9c-2b89-498e-becd-1e37cf111451 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.826042] env[62000]: DEBUG nova.policy [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '814f813a2b0f4a95b0496e9532696f48', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '542363220dde4ba4bbd40da8adbf5466', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 706.834097] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 706.834097] env[62000]: value = "task-881893" [ 706.834097] env[62000]: _type = "Task" [ 706.834097] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.843274] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881893, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 706.936432] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e662e4-8a29-3f6f-89b8-ca8ead986316, 'name': SearchDatastore_Task, 'duration_secs': 0.012749} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 706.936925] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 706.937243] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 706.937486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 706.937630] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 706.937806] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 706.938540] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e35bda2-e0e6-4079-a1e1-842568d75226 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.946409] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 706.946591] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 706.947346] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ac0121c-88cb-468f-9e5e-26e3ca5181c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.956233] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 706.956233] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52401895-6f64-1a1a-8c90-539852661de9" [ 706.956233] env[62000]: _type = "Task" [ 706.956233] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 706.963102] env[62000]: DEBUG nova.network.neutron [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.968435] env[62000]: DEBUG nova.network.neutron [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updating instance_info_cache with network_info: [{"id": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "address": "fa:16:3e:b3:ee:69", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d6933ed-ca", "ovs_interfaceid": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "address": "fa:16:3e:3b:be:31", "network": {"id": "9c93ba57-012b-4f3e-97a0-d21385b0f45b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-787006636", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.221", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap556236e8-eb", "ovs_interfaceid": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "10672d07-d10e-48f7-b383-87348134adb9", "address": "fa:16:3e:c2:43:eb", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10672d07-d1", "ovs_interfaceid": "10672d07-d10e-48f7-b383-87348134adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 706.970859] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52401895-6f64-1a1a-8c90-539852661de9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.023302] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Successfully updated port: c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 707.159136] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 707.180801] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Releasing lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.181178] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-plugged-10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.181312] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.181527] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.181710] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.181900] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] No waiting events found dispatching network-vif-plugged-10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 707.182109] env[62000]: WARNING nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received unexpected event network-vif-plugged-10672d07-d10e-48f7-b383-87348134adb9 for instance with vm_state building and task_state spawning. [ 707.183029] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-changed-10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.184250] env[62000]: DEBUG nova.compute.manager [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing instance network info cache due to event network-changed-10672d07-d10e-48f7-b383-87348134adb9. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.184464] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquiring lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.345622] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881893, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.379515] env[62000]: DEBUG nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Received event network-vif-deleted-1f4c2dd6-fe3a-40c6-aa27-0af8fba761b7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.379805] env[62000]: DEBUG nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Received event network-vif-plugged-c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.379954] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Acquiring lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.380332] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.380524] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.380692] env[62000]: DEBUG nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] No waiting events found dispatching network-vif-plugged-c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 707.380895] env[62000]: WARNING nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Received unexpected event network-vif-plugged-c0059cab-c384-49ca-b0fa-1e46163e688c for instance with vm_state building and task_state spawning. [ 707.382133] env[62000]: DEBUG nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Received event network-changed-c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 707.382133] env[62000]: DEBUG nova.compute.manager [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Refreshing instance network info cache due to event network-changed-c0059cab-c384-49ca-b0fa-1e46163e688c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 707.382133] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Acquiring lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.382133] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Acquired lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.382133] env[62000]: DEBUG nova.network.neutron [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Refreshing network info cache for port c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 707.472903] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 707.473647] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance network_info: |[{"id": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "address": "fa:16:3e:b3:ee:69", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d6933ed-ca", "ovs_interfaceid": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "address": "fa:16:3e:3b:be:31", "network": {"id": "9c93ba57-012b-4f3e-97a0-d21385b0f45b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-787006636", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.221", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap556236e8-eb", "ovs_interfaceid": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "10672d07-d10e-48f7-b383-87348134adb9", "address": "fa:16:3e:c2:43:eb", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10672d07-d1", "ovs_interfaceid": "10672d07-d10e-48f7-b383-87348134adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 707.474194] env[62000]: INFO nova.compute.manager [-] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Took 1.77 seconds to deallocate network for instance. [ 707.474864] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52401895-6f64-1a1a-8c90-539852661de9, 'name': SearchDatastore_Task, 'duration_secs': 0.009896} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.479168] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Acquired lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.479168] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Refreshing network info cache for port 10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 707.479595] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:ee:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3d6933ed-ca06-4521-a288-8c4c4cacefd1', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:3b:be:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3c29724c-5452-441a-8060-5bf89d1f5847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '556236e8-eb70-4dc1-8fbd-5090a7df2605', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:c2:43:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c9bc2632-36f9-4912-8782-8bbb789f909d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '10672d07-d10e-48f7-b383-87348134adb9', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 707.492963] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Creating folder: Project (e45b26d92ce540cea0f7cd27f52da2ec). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 707.497293] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Successfully created port: f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 707.502380] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cb120d31-dc2a-4159-819a-dff644417f74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.504727] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-82caae50-ffbe-4607-9483-bf0902e2aae8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.520583] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 707.520583] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5228ddd9-bdc0-164b-41ab-6e634abad972" [ 707.520583] env[62000]: _type = "Task" [ 707.520583] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.525619] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Created folder: Project (e45b26d92ce540cea0f7cd27f52da2ec) in parent group-v201431. [ 707.525829] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Creating folder: Instances. Parent ref: group-v201458. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 707.526624] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-52bc1e8a-82db-43f4-b69a-2ec4f23efdbd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.528890] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.532559] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5228ddd9-bdc0-164b-41ab-6e634abad972, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.535884] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Created folder: Instances in parent group-v201458. [ 707.536039] env[62000]: DEBUG oslo.service.loopingcall [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 707.536242] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 707.536444] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4f07ca6-a412-4658-8737-c3f627885d77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.565070] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 707.565070] env[62000]: value = "task-881896" [ 707.565070] env[62000]: _type = "Task" [ 707.565070] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.573720] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881896, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.602512] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed247e8b-1df1-421b-857a-1ea6dd795b19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.614317] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2c3393-1c21-41c0-8244-e8e4b4cee9f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.646342] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a11fc4d7-9173-4655-bc12-95234c73b61b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.654028] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa3641c1-ca42-48f4-b566-c0519042d167 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.673055] env[62000]: DEBUG nova.compute.provider_tree [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.844462] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881893, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.574445} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 707.845228] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 5e80a5c8-030c-4ad8-90c4-26136fa39d71/5e80a5c8-030c-4ad8-90c4-26136fa39d71.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 707.845509] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 707.845799] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1828c83-e2dc-4c28-8f76-a3352a407f9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.853530] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 707.853530] env[62000]: value = "task-881897" [ 707.853530] env[62000]: _type = "Task" [ 707.853530] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 707.862672] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 707.950154] env[62000]: DEBUG nova.network.neutron [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.966266] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updated VIF entry in instance network info cache for port 10672d07-d10e-48f7-b383-87348134adb9. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 707.967100] env[62000]: DEBUG nova.network.neutron [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updating instance_info_cache with network_info: [{"id": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "address": "fa:16:3e:b3:ee:69", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.121", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3d6933ed-ca", "ovs_interfaceid": "3d6933ed-ca06-4521-a288-8c4c4cacefd1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "address": "fa:16:3e:3b:be:31", "network": {"id": "9c93ba57-012b-4f3e-97a0-d21385b0f45b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-787006636", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.221", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3c29724c-5452-441a-8060-5bf89d1f5847", "external-id": "nsx-vlan-transportzone-683", "segmentation_id": 683, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap556236e8-eb", "ovs_interfaceid": "556236e8-eb70-4dc1-8fbd-5090a7df2605", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "10672d07-d10e-48f7-b383-87348134adb9", "address": "fa:16:3e:c2:43:eb", "network": {"id": "0c4c76d8-dedc-4276-b50f-2aaaf2e7ad80", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-2134208641", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.193", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c9bc2632-36f9-4912-8782-8bbb789f909d", "external-id": "nsx-vlan-transportzone-897", "segmentation_id": 897, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap10672d07-d1", "ovs_interfaceid": "10672d07-d10e-48f7-b383-87348134adb9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.016211] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.032718] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5228ddd9-bdc0-164b-41ab-6e634abad972, 'name': SearchDatastore_Task, 'duration_secs': 0.030951} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.033132] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.033421] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] eb27703f-b657-423a-90a9-a7c024a2e473/eb27703f-b657-423a-90a9-a7c024a2e473.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 708.033693] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f4c7b27-d039-4f00-93b7-cb9902e8b248 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.040034] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 708.040034] env[62000]: value = "task-881898" [ 708.040034] env[62000]: _type = "Task" [ 708.040034] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.050766] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.075337] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881896, 'name': CreateVM_Task, 'duration_secs': 0.427164} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.075520] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 708.076425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.076691] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.076943] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 708.077216] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c06d0b0-dab6-45ed-bb17-772323530ac0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.082390] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 708.082390] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5208aaff-6681-6c52-ad81-ad58973d3ed1" [ 708.082390] env[62000]: _type = "Task" [ 708.082390] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.090680] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5208aaff-6681-6c52-ad81-ad58973d3ed1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.128947] env[62000]: DEBUG nova.network.neutron [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.178422] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 708.183120] env[62000]: DEBUG nova.scheduler.client.report [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.206987] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 708.207503] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 708.207503] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 708.207721] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 708.207879] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 708.208066] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 708.208289] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 708.208450] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 708.208659] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 708.209496] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 708.209496] env[62000]: DEBUG nova.virt.hardware [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 708.210149] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1891e850-bd98-466f-9844-e35a66b225e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.218460] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ffa525-58dd-4846-83ec-a60717c11a27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.364042] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065557} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.364385] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 708.365229] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fe7db4d-3211-4f0e-abee-af7342f2315a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.386887] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] 5e80a5c8-030c-4ad8-90c4-26136fa39d71/5e80a5c8-030c-4ad8-90c4-26136fa39d71.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 708.387285] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f90c0509-919c-40f6-918e-7b213f556292 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.406936] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 708.406936] env[62000]: value = "task-881899" [ 708.406936] env[62000]: _type = "Task" [ 708.406936] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.416874] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.470305] env[62000]: DEBUG oslo_concurrency.lockutils [req-e85bf886-f59c-4e58-b8fd-4d065cf38024 req-55233bcc-c5ac-41c3-9291-fed24334f976 service nova] Releasing lock "refresh_cache-1d9b67b5-c0cf-41ff-a838-79caf8789609" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.551385] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881898, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.592544] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5208aaff-6681-6c52-ad81-ad58973d3ed1, 'name': SearchDatastore_Task, 'duration_secs': 0.00997} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.592863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.593172] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 708.593421] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 708.593574] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.593756] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 708.594028] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2d0bb495-78a1-4c9a-92bd-a8f2faa3eb30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.602020] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 708.602243] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 708.602952] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dd57679-5d51-452f-8337-8268e6e06425 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.608117] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 708.608117] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526d15cd-3e3f-cd69-19d1-74cc58c45763" [ 708.608117] env[62000]: _type = "Task" [ 708.608117] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.616246] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526d15cd-3e3f-cd69-19d1-74cc58c45763, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 708.632042] env[62000]: DEBUG oslo_concurrency.lockutils [req-8fefb262-c139-447c-bba0-6a88b67427f3 req-4c11fcca-2c1e-4131-9d84-a0130dbae291 service nova] Releasing lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.632518] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquired lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.632595] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 708.687177] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.541s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 708.687390] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 708.690050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.071s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.691467] env[62000]: INFO nova.compute.claims [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.917942] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881899, 'name': ReconfigVM_Task, 'duration_secs': 0.346858} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 708.918350] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Reconfigured VM instance instance-00000023 to attach disk [datastore2] 5e80a5c8-030c-4ad8-90c4-26136fa39d71/5e80a5c8-030c-4ad8-90c4-26136fa39d71.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 708.919054] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-524377e7-30f3-403a-91fa-9a4e17e8241d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.926540] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 708.926540] env[62000]: value = "task-881900" [ 708.926540] env[62000]: _type = "Task" [ 708.926540] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 708.934947] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881900, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.050544] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881898, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.119057] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526d15cd-3e3f-cd69-19d1-74cc58c45763, 'name': SearchDatastore_Task, 'duration_secs': 0.009695} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.119581] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3542b3fa-9dec-4594-b421-0a53bb958c3a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.125023] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 709.125023] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ea370e-3179-d748-dbab-35b0f91723de" [ 709.125023] env[62000]: _type = "Task" [ 709.125023] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.132817] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ea370e-3179-d748-dbab-35b0f91723de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.187681] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 709.196606] env[62000]: DEBUG nova.compute.utils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 709.200327] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 709.200719] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 709.331168] env[62000]: DEBUG nova.policy [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3018dc2fa8d413aa464bf4e0c32341e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d131f95ad3949d89cd6f36f6648d3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 709.437088] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881900, 'name': Rename_Task, 'duration_secs': 0.139376} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.438476] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 709.438476] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-235c907b-64cd-42b9-9c55-f23bc96ce7bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.443957] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 709.443957] env[62000]: value = "task-881901" [ 709.443957] env[62000]: _type = "Task" [ 709.443957] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.452180] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.558817] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881898, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.576538] env[62000]: DEBUG nova.network.neutron [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updating instance_info_cache with network_info: [{"id": "c0059cab-c384-49ca-b0fa-1e46163e688c", "address": "fa:16:3e:89:bf:85", "network": {"id": "c4fade47-cb22-4277-86be-849d1fceb2b2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1349651934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76de44563cdd4a3883d5153555c2e48e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0059cab-c3", "ovs_interfaceid": "c0059cab-c384-49ca-b0fa-1e46163e688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.635254] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ea370e-3179-d748-dbab-35b0f91723de, 'name': SearchDatastore_Task, 'duration_secs': 0.0115} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.635537] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.635797] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 1d9b67b5-c0cf-41ff-a838-79caf8789609/1d9b67b5-c0cf-41ff-a838-79caf8789609.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 709.636078] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-47be5f83-46f7-44ea-96ea-436ee1f9703d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.643698] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 709.643698] env[62000]: value = "task-881902" [ 709.643698] env[62000]: _type = "Task" [ 709.643698] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 709.653119] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881902, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 709.701725] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 709.868596] env[62000]: DEBUG nova.compute.manager [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Received event network-vif-plugged-f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 709.869081] env[62000]: DEBUG oslo_concurrency.lockutils [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] Acquiring lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.869941] env[62000]: DEBUG oslo_concurrency.lockutils [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.870199] env[62000]: DEBUG oslo_concurrency.lockutils [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.870527] env[62000]: DEBUG nova.compute.manager [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] No waiting events found dispatching network-vif-plugged-f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 709.870848] env[62000]: WARNING nova.compute.manager [req-338bcd66-41f9-403e-9062-5ca0bea242ca req-1ae67078-3aa9-45b1-822c-df562b06448e service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Received unexpected event network-vif-plugged-f077ad17-a015-4c19-aa71-f1f595d55c65 for instance with vm_state building and task_state spawning. [ 709.939478] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Successfully created port: a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 709.956040] env[62000]: DEBUG oslo_vmware.api [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881901, 'name': PowerOnVM_Task, 'duration_secs': 0.431844} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 709.956227] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 709.956867] env[62000]: INFO nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Took 9.15 seconds to spawn the instance on the hypervisor. [ 709.956867] env[62000]: DEBUG nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 709.957460] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2879989e-8ea6-412a-88bc-77e84ae77631 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.055360] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881898, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.537509} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.055957] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] eb27703f-b657-423a-90a9-a7c024a2e473/eb27703f-b657-423a-90a9-a7c024a2e473.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 710.055957] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.056142] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8de592a2-a19e-4729-a07a-473dc9c5e51d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.063237] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 710.063237] env[62000]: value = "task-881903" [ 710.063237] env[62000]: _type = "Task" [ 710.063237] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.072556] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881903, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.074704] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5bc11dd-e4f4-41d9-9622-d9aa2d94ed3e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.080349] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Releasing lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.080775] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance network_info: |[{"id": "c0059cab-c384-49ca-b0fa-1e46163e688c", "address": "fa:16:3e:89:bf:85", "network": {"id": "c4fade47-cb22-4277-86be-849d1fceb2b2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1349651934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76de44563cdd4a3883d5153555c2e48e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0059cab-c3", "ovs_interfaceid": "c0059cab-c384-49ca-b0fa-1e46163e688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 710.083688] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:bf:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f847601f-7479-48eb-842f-41f94eea8537', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c0059cab-c384-49ca-b0fa-1e46163e688c', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 710.092188] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Creating folder: Project (76de44563cdd4a3883d5153555c2e48e). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.092617] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c149c254-d302-4493-9b35-25077f6282d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.095516] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203edfdf-a6cf-4418-a914-48a55219cd2e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.130903] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fcd3dbf-fb28-4c8a-9e7f-c69d0fadc66f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.133824] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Created folder: Project (76de44563cdd4a3883d5153555c2e48e) in parent group-v201431. [ 710.134056] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Creating folder: Instances. Parent ref: group-v201461. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 710.134337] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-692f333f-19cc-4f81-8c0b-0a17e61475db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.141867] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f889ae3-1645-4b90-89a8-e1060932c16c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.147832] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Created folder: Instances in parent group-v201461. [ 710.148108] env[62000]: DEBUG oslo.service.loopingcall [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 710.151999] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 710.152374] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1e341ad1-d933-4b06-b279-88217180c2e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.179273] env[62000]: DEBUG nova.compute.provider_tree [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.184916] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881902, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.194050] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 710.194050] env[62000]: value = "task-881906" [ 710.194050] env[62000]: _type = "Task" [ 710.194050] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.203101] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881906, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.203934] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Successfully updated port: f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 710.333468] env[62000]: DEBUG nova.compute.manager [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Received event network-changed-f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 710.333722] env[62000]: DEBUG nova.compute.manager [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Refreshing instance network info cache due to event network-changed-f077ad17-a015-4c19-aa71-f1f595d55c65. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 710.333892] env[62000]: DEBUG oslo_concurrency.lockutils [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] Acquiring lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.334054] env[62000]: DEBUG oslo_concurrency.lockutils [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] Acquired lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.334219] env[62000]: DEBUG nova.network.neutron [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Refreshing network info cache for port f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 710.457496] env[62000]: DEBUG nova.compute.manager [None req-5290924f-eb27-4523-8f7d-7e8562bdd4af tempest-ServerDiagnosticsV248Test-229552800 tempest-ServerDiagnosticsV248Test-229552800-project-admin] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 710.458652] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d5c7e27-c9dc-4c10-a397-2968c8ad20d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.465675] env[62000]: INFO nova.compute.manager [None req-5290924f-eb27-4523-8f7d-7e8562bdd4af tempest-ServerDiagnosticsV248Test-229552800 tempest-ServerDiagnosticsV248Test-229552800-project-admin] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Retrieving diagnostics [ 710.466650] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3dfd5cc-6b1c-4f7d-92d9-eb9ce48e2b9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.502589] env[62000]: INFO nova.compute.manager [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Took 42.68 seconds to build instance. [ 710.573158] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881903, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084598} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.573484] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 710.574239] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b82436d-176a-49e2-9f2c-f02bf5b63e8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.596354] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] eb27703f-b657-423a-90a9-a7c024a2e473/eb27703f-b657-423a-90a9-a7c024a2e473.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 710.596674] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44ed0ec1-0ad4-4519-b8ca-a537d16819e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.617716] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 710.617716] env[62000]: value = "task-881907" [ 710.617716] env[62000]: _type = "Task" [ 710.617716] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.628414] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881907, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.656317] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881902, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.548427} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 710.656583] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 1d9b67b5-c0cf-41ff-a838-79caf8789609/1d9b67b5-c0cf-41ff-a838-79caf8789609.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 710.656803] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 710.657069] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ab6976f7-dafd-4261-ada3-1aef04a0978e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.663484] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 710.663484] env[62000]: value = "task-881908" [ 710.663484] env[62000]: _type = "Task" [ 710.663484] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 710.671349] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.688577] env[62000]: DEBUG nova.scheduler.client.report [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.704922] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881906, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 710.709724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 710.718052] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 710.743866] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 710.744149] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 710.744313] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 710.744496] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 710.744647] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 710.744822] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 710.745061] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 710.745184] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 710.745352] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 710.745520] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 710.746029] env[62000]: DEBUG nova.virt.hardware [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 710.746655] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0333465-f6bd-4a3d-9e4b-594f5686fc7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.754596] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcc5d22-56d8-4fa2-8add-37823f413ee5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.879758] env[62000]: DEBUG nova.network.neutron [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.935225] env[62000]: DEBUG nova.network.neutron [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.006613] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7eacd126-3b53-4fe6-ab6b-58d2b3b9629b tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.289s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.131767] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881907, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.173270] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08622} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.173442] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 711.174354] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c590d13-f7d8-4719-b4d4-92922da09829 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.193465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.194081] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 711.206120] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Reconfiguring VM instance instance-00000022 to attach disk [datastore2] 1d9b67b5-c0cf-41ff-a838-79caf8789609/1d9b67b5-c0cf-41ff-a838-79caf8789609.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 711.206120] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.601s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.207973] env[62000]: INFO nova.compute.claims [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 711.211393] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb73cb65-c12e-42f4-89fe-32ffc0da66e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.240215] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881906, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.240369] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 711.240369] env[62000]: value = "task-881909" [ 711.240369] env[62000]: _type = "Task" [ 711.240369] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.249217] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881909, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.440349] env[62000]: DEBUG oslo_concurrency.lockutils [req-c0de9955-ff15-4caa-90b0-570e9cbf09c7 req-e5714426-b59f-4655-8e2d-9ca2b54f081b service nova] Releasing lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.440880] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquired lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.441098] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.509060] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.638392] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881907, 'name': ReconfigVM_Task, 'duration_secs': 0.639731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.638961] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfigured VM instance instance-00000024 to attach disk [datastore2] eb27703f-b657-423a-90a9-a7c024a2e473/eb27703f-b657-423a-90a9-a7c024a2e473.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 711.640172] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a5b5cd5b-4416-42a2-b2e1-ce334f9fa718 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.647611] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 711.647611] env[62000]: value = "task-881910" [ 711.647611] env[62000]: _type = "Task" [ 711.647611] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.656908] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881910, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.709571] env[62000]: DEBUG nova.compute.utils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 711.712754] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 711.712754] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 711.722370] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881906, 'name': CreateVM_Task, 'duration_secs': 1.261878} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 711.723165] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 711.723867] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.724043] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.724374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 711.724897] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e95823d5-5bb0-4867-948c-a2a0108bb40a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.734040] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 711.734040] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e15a58-e7c9-98b0-bd09-343764665e92" [ 711.734040] env[62000]: _type = "Task" [ 711.734040] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 711.742915] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e15a58-e7c9-98b0-bd09-343764665e92, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.746684] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Successfully updated port: a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 711.753646] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881909, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 711.785639] env[62000]: DEBUG nova.policy [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a78e29b835a14ca89dc766c97cf4bffc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ea8a6296044141e38bcefc93f1718657', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 711.990596] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.043917] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.158404] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881910, 'name': Rename_Task, 'duration_secs': 0.163146} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.158688] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 712.158945] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8a6b1cf-2d6a-4119-ba4f-47d3d38f14eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.165796] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 712.165796] env[62000]: value = "task-881911" [ 712.165796] env[62000]: _type = "Task" [ 712.165796] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.173815] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.214794] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 712.253426] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.253624] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.253937] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 712.259889] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e15a58-e7c9-98b0-bd09-343764665e92, 'name': SearchDatastore_Task, 'duration_secs': 0.012266} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.263552] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.263823] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 712.264086] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.264367] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.264450] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 712.269530] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16c7317a-1c80-48ee-861c-e23e077cb070 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.271677] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881909, 'name': ReconfigVM_Task, 'duration_secs': 0.519796} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.276320] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Reconfigured VM instance instance-00000022 to attach disk [datastore2] 1d9b67b5-c0cf-41ff-a838-79caf8789609/1d9b67b5-c0cf-41ff-a838-79caf8789609.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 712.281486] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3438678f-2ab4-43cc-824c-877d209f392e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.284228] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 712.284481] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 712.285741] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06c7ff15-488f-4e9b-8b5b-4af4d09a5c44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.290740] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 712.290740] env[62000]: value = "task-881912" [ 712.290740] env[62000]: _type = "Task" [ 712.290740] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.298746] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 712.298746] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522a21b4-fb59-4a64-9c81-437f37705246" [ 712.298746] env[62000]: _type = "Task" [ 712.298746] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.306057] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881912, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.311737] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522a21b4-fb59-4a64-9c81-437f37705246, 'name': SearchDatastore_Task, 'duration_secs': 0.011905} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.312908] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bfc820ac-96f0-4697-8ffb-15028e88a3db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.322437] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 712.322437] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52942c8c-e575-1f7e-c7da-fe00ce5060ca" [ 712.322437] env[62000]: _type = "Task" [ 712.322437] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.333961] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52942c8c-e575-1f7e-c7da-fe00ce5060ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.508658] env[62000]: DEBUG nova.network.neutron [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Updating instance_info_cache with network_info: [{"id": "f077ad17-a015-4c19-aa71-f1f595d55c65", "address": "fa:16:3e:f9:30:a4", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.46", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf077ad17-a0", "ovs_interfaceid": "f077ad17-a015-4c19-aa71-f1f595d55c65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.528137] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Successfully created port: b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 712.620270] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82e08818-56d4-4154-95a1-def5253a4259 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.628484] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ba0e5d-785a-4cc8-8f8d-37d23d09f7b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.663251] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e394304-cb3c-4899-ad16-eabb0db9eb76 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.672891] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6730048b-04b4-4e2a-8d88-d41c423116c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.681412] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881911, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.686569] env[62000]: DEBUG nova.compute.manager [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Received event network-vif-plugged-a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.686569] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.686569] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.686569] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.686569] env[62000]: DEBUG nova.compute.manager [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] No waiting events found dispatching network-vif-plugged-a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 712.686569] env[62000]: WARNING nova.compute.manager [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Received unexpected event network-vif-plugged-a40f58be-07f1-43ce-afd9-e0ea0c78b634 for instance with vm_state building and task_state spawning. [ 712.686569] env[62000]: DEBUG nova.compute.manager [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Received event network-changed-a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 712.686569] env[62000]: DEBUG nova.compute.manager [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Refreshing instance network info cache due to event network-changed-a40f58be-07f1-43ce-afd9-e0ea0c78b634. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 712.686569] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.694872] env[62000]: DEBUG nova.compute.provider_tree [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.808716] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881912, 'name': Rename_Task, 'duration_secs': 0.162323} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.809181] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 712.809300] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c54bd8b5-758b-403d-b04a-a7116d13e2ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.816551] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 712.816551] env[62000]: value = "task-881913" [ 712.816551] env[62000]: _type = "Task" [ 712.816551] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.830733] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 712.835304] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52942c8c-e575-1f7e-c7da-fe00ce5060ca, 'name': SearchDatastore_Task, 'duration_secs': 0.012125} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 712.836157] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 712.838193] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.838469] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 49f6c198-13b7-4c07-81d8-c010e7b0598e/49f6c198-13b7-4c07-81d8-c010e7b0598e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 712.839048] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c38c8b0b-6ece-4d1b-8feb-dd5211d2a331 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.846422] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 712.846422] env[62000]: value = "task-881914" [ 712.846422] env[62000]: _type = "Task" [ 712.846422] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 712.855861] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881914, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.017458] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Releasing lock "refresh_cache-415cc4a5-7610-4678-971d-cd00a0e8b54d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.017458] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance network_info: |[{"id": "f077ad17-a015-4c19-aa71-f1f595d55c65", "address": "fa:16:3e:f9:30:a4", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.46", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf077ad17-a0", "ovs_interfaceid": "f077ad17-a015-4c19-aa71-f1f595d55c65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 713.017458] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f9:30:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f077ad17-a015-4c19-aa71-f1f595d55c65', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.026025] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Creating folder: Project (542363220dde4ba4bbd40da8adbf5466). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.026025] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65fa55ad-138f-4d71-a829-e34ee51e069c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.039460] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Created folder: Project (542363220dde4ba4bbd40da8adbf5466) in parent group-v201431. [ 713.039460] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Creating folder: Instances. Parent ref: group-v201464. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.039599] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f492f87d-a6ea-4781-a23c-923232265ba5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.051019] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Created folder: Instances in parent group-v201464. [ 713.051019] env[62000]: DEBUG oslo.service.loopingcall [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.051019] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 713.051019] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-87d5cd50-1ca9-48b4-8b46-2fc30523507c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.077019] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.077019] env[62000]: value = "task-881917" [ 713.077019] env[62000]: _type = "Task" [ 713.077019] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.085515] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881917, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.086588] env[62000]: DEBUG nova.network.neutron [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.181457] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881911, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.198077] env[62000]: DEBUG nova.scheduler.client.report [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 713.227702] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 713.254402] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 713.254783] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 713.255064] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 713.255443] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 713.255725] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 713.255998] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 713.256398] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 713.256681] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 713.257030] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 713.257367] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 713.257634] env[62000]: DEBUG nova.virt.hardware [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 713.258814] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781f50b8-6906-4def-acbb-9245efd36b43 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.268145] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2ea0e0-593a-4169-9cbf-5997de47e540 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.327361] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.356990] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881914, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.587721] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881917, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.590041] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.590041] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Instance network_info: |[{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 713.590247] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.590351] env[62000]: DEBUG nova.network.neutron [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Refreshing network info cache for port a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 713.591597] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:00:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a40f58be-07f1-43ce-afd9-e0ea0c78b634', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 713.598888] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating folder: Project (4d131f95ad3949d89cd6f36f6648d3f1). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.602643] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d50bf99b-9cf8-4aec-90ac-0d6d49adb3c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.614339] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created folder: Project (4d131f95ad3949d89cd6f36f6648d3f1) in parent group-v201431. [ 713.614612] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating folder: Instances. Parent ref: group-v201467. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 713.614861] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eb748a12-8832-48c9-980c-289f2aabd0fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.624308] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created folder: Instances in parent group-v201467. [ 713.624566] env[62000]: DEBUG oslo.service.loopingcall [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.624765] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 713.624974] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f4f206d3-5140-4ce2-a2be-ab0c3a03f329 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.645559] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 713.645559] env[62000]: value = "task-881920" [ 713.645559] env[62000]: _type = "Task" [ 713.645559] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.653698] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881920, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.676528] env[62000]: DEBUG oslo_vmware.api [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-881911, 'name': PowerOnVM_Task, 'duration_secs': 1.147222} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.676832] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 713.677124] env[62000]: INFO nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Took 10.42 seconds to spawn the instance on the hypervisor. [ 713.677208] env[62000]: DEBUG nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 713.677969] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e428416c-fa77-4a40-99f3-43ed5218a0db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.703578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.704158] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 713.706656] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.177s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.706936] env[62000]: DEBUG nova.objects.instance [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lazy-loading 'resources' on Instance uuid 1419e5a8-a78f-44f1-bfc4-00fcf176bf35 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 713.827605] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.858448] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881914, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541176} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 713.859080] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 49f6c198-13b7-4c07-81d8-c010e7b0598e/49f6c198-13b7-4c07-81d8-c010e7b0598e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 713.859080] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 713.859317] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-33877922-e840-4dc7-9d34-fb75aac0a3b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.867729] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 713.867729] env[62000]: value = "task-881921" [ 713.867729] env[62000]: _type = "Task" [ 713.867729] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 713.875674] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881921, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 713.903049] env[62000]: DEBUG nova.network.neutron [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updated VIF entry in instance network info cache for port a40f58be-07f1-43ce-afd9-e0ea0c78b634. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 713.903502] env[62000]: DEBUG nova.network.neutron [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.087807] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881917, 'name': CreateVM_Task, 'duration_secs': 0.581684} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.087985] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 714.088867] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.089126] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.089503] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 714.089835] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe9c54f9-6f7e-4b24-86a0-bd8e13063387 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.094644] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 714.094644] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52891190-5dda-a69e-ba1d-895dc7105309" [ 714.094644] env[62000]: _type = "Task" [ 714.094644] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.103379] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52891190-5dda-a69e-ba1d-895dc7105309, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.155951] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881920, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.197552] env[62000]: INFO nova.compute.manager [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Took 44.53 seconds to build instance. [ 714.210402] env[62000]: DEBUG nova.compute.utils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 714.212561] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 714.212991] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 714.303754] env[62000]: DEBUG nova.policy [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd154052032f342a195170e81d4bbabc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c051bb06d4c04417902263736e94a8b0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 714.334449] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881913, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.381124] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881921, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065877} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.381331] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 714.382214] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa7d94ab-ef5a-45c5-94ec-c4119755fc55 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.412672] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Reconfiguring VM instance instance-00000025 to attach disk [datastore1] 49f6c198-13b7-4c07-81d8-c010e7b0598e/49f6c198-13b7-4c07-81d8-c010e7b0598e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 714.414091] env[62000]: DEBUG oslo_concurrency.lockutils [req-f5859dcd-b516-40a6-989a-acde52c1ddc2 req-a727c83e-9352-40bd-a252-22c39c8e558e service nova] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.414556] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cc9ca29-808d-4351-ad31-55366baeeaa9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.438040] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 714.438040] env[62000]: value = "task-881922" [ 714.438040] env[62000]: _type = "Task" [ 714.438040] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.447652] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881922, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.519897] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Successfully updated port: b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 714.567987] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 714.568943] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-227979bd-5d00-47fb-8421-4102fb4a3e5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.578140] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 714.578340] env[62000]: ERROR oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk due to incomplete transfer. [ 714.578577] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-93e2f68d-ca71-47fb-a832-164449ea8a28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.585535] env[62000]: DEBUG oslo_vmware.rw_handles [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5273d76d-46be-303b-9ded-c2c67ef6d09f/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 714.585808] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Uploaded image 3f9a0b46-aee3-432d-9f8a-fbc1a16bd49e to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 714.587887] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 714.590492] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9eda2d4e-9b7f-42b6-8981-5c658cc3b9a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.597062] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 714.597062] env[62000]: value = "task-881923" [ 714.597062] env[62000]: _type = "Task" [ 714.597062] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.611796] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52891190-5dda-a69e-ba1d-895dc7105309, 'name': SearchDatastore_Task, 'duration_secs': 0.021307} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.614950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.615239] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 714.615565] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.615777] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.616255] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 714.616255] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881923, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.616464] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5467332c-4ef5-41f1-aeec-7df2e0327ea7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.624437] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 714.624667] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 714.625348] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-750ceb9f-49e7-47d8-85d0-f51e2b8e4420 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.634129] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 714.634129] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52834527-7178-79cf-1f61-46fd5a29cafc" [ 714.634129] env[62000]: _type = "Task" [ 714.634129] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.644273] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52834527-7178-79cf-1f61-46fd5a29cafc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.654397] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881920, 'name': CreateVM_Task, 'duration_secs': 0.610226} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.654574] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 714.655269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.655453] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.655871] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 714.656022] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc6b4f8d-de8e-4186-b721-8115340ab58a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.660455] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 714.660455] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216eb07-bbdb-2705-602e-486cbcdac347" [ 714.660455] env[62000]: _type = "Task" [ 714.660455] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 714.670638] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216eb07-bbdb-2705-602e-486cbcdac347, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 714.690163] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d589b7-e5c1-464f-a87d-a5093d18bd01 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.696313] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea16d57-ed6a-4977-ba47-c13c7d7fac10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.699991] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3f7656f9-62f4-4741-9138-20411383f33e tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 131.934s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.730165] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 714.734567] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b94ecaa0-5ca8-4f11-b026-51ae1c80e569 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.741035] env[62000]: DEBUG nova.compute.manager [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Received event network-vif-plugged-b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.741035] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Acquiring lock "72688992-333f-459d-9d05-f7c728961a6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.741035] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Lock "72688992-333f-459d-9d05-f7c728961a6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.741035] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Lock "72688992-333f-459d-9d05-f7c728961a6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.741035] env[62000]: DEBUG nova.compute.manager [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] No waiting events found dispatching network-vif-plugged-b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 714.741035] env[62000]: WARNING nova.compute.manager [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Received unexpected event network-vif-plugged-b63c8316-148a-47a4-b17a-d394a10a072c for instance with vm_state building and task_state spawning. [ 714.741035] env[62000]: DEBUG nova.compute.manager [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Received event network-changed-b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 714.741035] env[62000]: DEBUG nova.compute.manager [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Refreshing instance network info cache due to event network-changed-b63c8316-148a-47a4-b17a-d394a10a072c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 714.741035] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Acquiring lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 714.741480] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Acquired lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 714.741480] env[62000]: DEBUG nova.network.neutron [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Refreshing network info cache for port b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 714.749727] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5df4f8c2-7bd2-4aaa-b3aa-d89e96a34ea4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.754572] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Successfully created port: 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 714.769120] env[62000]: DEBUG nova.compute.provider_tree [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 714.828035] env[62000]: DEBUG oslo_vmware.api [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881913, 'name': PowerOnVM_Task, 'duration_secs': 1.602354} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 714.828336] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 714.828572] env[62000]: INFO nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Took 16.49 seconds to spawn the instance on the hypervisor. [ 714.828771] env[62000]: DEBUG nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 714.829605] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f0057b9-0eeb-47aa-b054-3f18c35de235 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.950214] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881922, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.024106] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.110807] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881923, 'name': Destroy_Task, 'duration_secs': 0.297971} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.111033] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Destroyed the VM [ 715.111369] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 715.111629] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b40a3b6f-8765-4f3f-ac60-410d952916d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.121868] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 715.121868] env[62000]: value = "task-881924" [ 715.121868] env[62000]: _type = "Task" [ 715.121868] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.131950] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881924, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.148949] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52834527-7178-79cf-1f61-46fd5a29cafc, 'name': SearchDatastore_Task, 'duration_secs': 0.009404} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.149796] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ba9dafe-10ff-4681-bb23-e7c7df6359ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.156121] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 715.156121] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ff81d-7e5f-56b1-54c0-7fb6b3d967df" [ 715.156121] env[62000]: _type = "Task" [ 715.156121] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.164596] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ff81d-7e5f-56b1-54c0-7fb6b3d967df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.172894] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216eb07-bbdb-2705-602e-486cbcdac347, 'name': SearchDatastore_Task, 'duration_secs': 0.008946} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.173255] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.173512] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 715.173741] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.202572] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 715.287820] env[62000]: DEBUG nova.network.neutron [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 715.295249] env[62000]: ERROR nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] [req-5b34b180-eb0b-40ff-a59b-a128d30b3c02] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-5b34b180-eb0b-40ff-a59b-a128d30b3c02"}]} [ 715.321294] env[62000]: DEBUG nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 715.340838] env[62000]: DEBUG nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 715.341710] env[62000]: DEBUG nova.compute.provider_tree [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 715.356716] env[62000]: INFO nova.compute.manager [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Took 49.57 seconds to build instance. [ 715.374645] env[62000]: DEBUG nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 715.403024] env[62000]: DEBUG nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 715.430550] env[62000]: DEBUG nova.network.neutron [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.450440] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881922, 'name': ReconfigVM_Task, 'duration_secs': 0.608163} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.450865] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Reconfigured VM instance instance-00000025 to attach disk [datastore1] 49f6c198-13b7-4c07-81d8-c010e7b0598e/49f6c198-13b7-4c07-81d8-c010e7b0598e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 715.452723] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d95a9d3-82ed-455b-a524-681e17936970 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.460273] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 715.460273] env[62000]: value = "task-881925" [ 715.460273] env[62000]: _type = "Task" [ 715.460273] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.476451] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881925, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.566593] env[62000]: DEBUG nova.compute.manager [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Received event network-changed-d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 715.566809] env[62000]: DEBUG nova.compute.manager [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Refreshing instance network info cache due to event network-changed-d725993d-bf92-4c34-9f24-1c0ed6557610. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 715.567034] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] Acquiring lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 715.567187] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] Acquired lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.567397] env[62000]: DEBUG nova.network.neutron [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Refreshing network info cache for port d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 715.632397] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881924, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.678164] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ff81d-7e5f-56b1-54c0-7fb6b3d967df, 'name': SearchDatastore_Task, 'duration_secs': 0.01008} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 715.678164] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.678164] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 415cc4a5-7610-4678-971d-cd00a0e8b54d/415cc4a5-7610-4678-971d-cd00a0e8b54d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 715.678164] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.678164] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 715.678344] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e5346437-bfef-4269-b93b-21ad0ed4e01e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.680187] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-566dcfc3-9d54-4d48-88a3-d55551bc2119 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.688605] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 715.688605] env[62000]: value = "task-881926" [ 715.688605] env[62000]: _type = "Task" [ 715.688605] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.689832] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 715.690021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 715.696035] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b1dd21e-02bc-4da6-b00a-ef094a851f34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.710057] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881926, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.710951] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 715.710951] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521fb1a4-8dc1-9ab5-1a7a-b21048b0d483" [ 715.710951] env[62000]: _type = "Task" [ 715.710951] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 715.726656] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521fb1a4-8dc1-9ab5-1a7a-b21048b0d483, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 715.727922] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.750790] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 715.779384] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 715.779577] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 715.779735] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 715.779956] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 715.780329] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 715.783402] env[62000]: DEBUG nova.virt.hardware [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 715.783402] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41269401-50b4-412d-a20e-5a2a40b533af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.795212] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b9fb61-be5b-4ae0-b9fe-ae1701e6d33e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.802176] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6dce30-2a82-4c50-aedf-88a2ce53f46c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.822351] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea31eb82-4f40-41cc-b320-9264afeafe54 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.857413] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d4594b8-caae-41bf-916b-d9a85d7f6b9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.860795] env[62000]: DEBUG oslo_concurrency.lockutils [None req-23be01e3-e497-49d8-a136-41815a4ea8dd tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 135.519s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.868126] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b52abb-ecf8-489f-a4e8-f58c2387819d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.888936] env[62000]: DEBUG nova.compute.provider_tree [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 715.933216] env[62000]: DEBUG oslo_concurrency.lockutils [req-3acb01ce-9960-40f5-a73a-b046b9ae553f req-c12333d1-a042-408d-999d-6a31d4e4282d service nova] Releasing lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 715.933674] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquired lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 715.933879] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 715.971404] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881925, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.133412] env[62000]: DEBUG oslo_vmware.api [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881924, 'name': RemoveSnapshot_Task, 'duration_secs': 0.649927} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.133775] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 716.134067] env[62000]: INFO nova.compute.manager [None req-ceba3d8c-c97d-4624-ac77-5cfdb567d9e7 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 14.37 seconds to snapshot the instance on the hypervisor. [ 716.199467] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881926, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.490959} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.199818] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 415cc4a5-7610-4678-971d-cd00a0e8b54d/415cc4a5-7610-4678-971d-cd00a0e8b54d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 716.200122] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 716.200424] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-02b33da4-40bf-44db-8911-963f99137aab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.207845] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 716.207845] env[62000]: value = "task-881927" [ 716.207845] env[62000]: _type = "Task" [ 716.207845] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.216547] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881927, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.225236] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521fb1a4-8dc1-9ab5-1a7a-b21048b0d483, 'name': SearchDatastore_Task, 'duration_secs': 0.022494} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.226044] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68ab37de-cc64-4d73-96df-951a2fd23928 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.231434] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 716.231434] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b5943b-3589-5551-fb8e-0e0c7f47ac72" [ 716.231434] env[62000]: _type = "Task" [ 716.231434] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.242172] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b5943b-3589-5551-fb8e-0e0c7f47ac72, 'name': SearchDatastore_Task, 'duration_secs': 0.008245} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.242509] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 716.242827] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 716.243153] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-833ce431-4206-42b9-816f-3f1df67d5bc3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.250134] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 716.250134] env[62000]: value = "task-881928" [ 716.250134] env[62000]: _type = "Task" [ 716.250134] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.258155] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881928, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.363491] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.394785] env[62000]: DEBUG nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.475607] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881925, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.533847] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 716.662200] env[62000]: DEBUG nova.network.neutron [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updated VIF entry in instance network info cache for port d725993d-bf92-4c34-9f24-1c0ed6557610. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 716.662655] env[62000]: DEBUG nova.network.neutron [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating instance_info_cache with network_info: [{"id": "d725993d-bf92-4c34-9f24-1c0ed6557610", "address": "fa:16:3e:d1:f7:5d", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.189", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd725993d-bf", "ovs_interfaceid": "d725993d-bf92-4c34-9f24-1c0ed6557610", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.717600] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881927, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.088731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 716.717902] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 716.718634] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3181bc77-11ff-473d-8f72-169a95a27b38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.743213] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Reconfiguring VM instance instance-00000026 to attach disk [datastore1] 415cc4a5-7610-4678-971d-cd00a0e8b54d/415cc4a5-7610-4678-971d-cd00a0e8b54d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 716.743522] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3e535774-d5b6-4c23-a198-0b23c25ea23d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.767535] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881928, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.768793] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 716.768793] env[62000]: value = "task-881929" [ 716.768793] env[62000]: _type = "Task" [ 716.768793] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.776570] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881929, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.857260] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Successfully updated port: 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 716.871109] env[62000]: DEBUG nova.network.neutron [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Updating instance_info_cache with network_info: [{"id": "b63c8316-148a-47a4-b17a-d394a10a072c", "address": "fa:16:3e:0b:93:8e", "network": {"id": "9d1b9150-5613-4b84-bcd3-bc87ed1cf28c", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-564908060-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea8a6296044141e38bcefc93f1718657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb63c8316-14", "ovs_interfaceid": "b63c8316-148a-47a4-b17a-d394a10a072c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 716.901128] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.194s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.905867] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.906199] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.209s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.907869] env[62000]: INFO nova.compute.claims [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 716.912620] env[62000]: DEBUG nova.compute.manager [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-vif-plugged-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 716.913520] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] Acquiring lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.913520] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.913520] env[62000]: DEBUG oslo_concurrency.lockutils [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.913520] env[62000]: DEBUG nova.compute.manager [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] No waiting events found dispatching network-vif-plugged-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 716.913774] env[62000]: WARNING nova.compute.manager [req-9c86b100-984e-47d3-ad26-e4c6130e050c req-d20d0f32-ec9e-466f-8342-20d6157995a6 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received unexpected event network-vif-plugged-1d7778e6-8a84-447d-a241-95ea69ab68ba for instance with vm_state building and task_state spawning. [ 716.920774] env[62000]: INFO nova.scheduler.client.report [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Deleted allocations for instance 1419e5a8-a78f-44f1-bfc4-00fcf176bf35 [ 716.942031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.942148] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.945018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.945018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.945018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.945018] env[62000]: INFO nova.compute.manager [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Terminating instance [ 716.948031] env[62000]: DEBUG nova.compute.manager [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 716.948031] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 716.949070] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f466081-b277-4b14-ab5f-1237d31499ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.957824] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 716.958080] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e62e015f-473e-491d-bddc-82bea73188d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.967229] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 716.967229] env[62000]: value = "task-881930" [ 716.967229] env[62000]: _type = "Task" [ 716.967229] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 716.975353] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881925, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 716.980109] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.166250] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5cfe9a6-f36a-4429-a4e2-c0939d4a569a req-496a644f-dfa1-40b4-a344-b9ad326f6767 service nova] Releasing lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.268885] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881928, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573219} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.269038] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 717.269588] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 717.270705] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3622e2a4-69cd-4564-a98a-5cb4264f2252 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.281582] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881929, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.283014] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 717.283014] env[62000]: value = "task-881931" [ 717.283014] env[62000]: _type = "Task" [ 717.283014] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.290269] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881931, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.366167] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.366353] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.366514] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.374924] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Releasing lock "refresh_cache-72688992-333f-459d-9d05-f7c728961a6d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.374924] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance network_info: |[{"id": "b63c8316-148a-47a4-b17a-d394a10a072c", "address": "fa:16:3e:0b:93:8e", "network": {"id": "9d1b9150-5613-4b84-bcd3-bc87ed1cf28c", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-564908060-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ea8a6296044141e38bcefc93f1718657", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "06cc7c49-c46c-4c1e-bf51-77e9ea802c40", "external-id": "nsx-vlan-transportzone-450", "segmentation_id": 450, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb63c8316-14", "ovs_interfaceid": "b63c8316-148a-47a4-b17a-d394a10a072c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 717.375538] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:93:8e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '06cc7c49-c46c-4c1e-bf51-77e9ea802c40', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b63c8316-148a-47a4-b17a-d394a10a072c', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 717.383536] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Creating folder: Project (ea8a6296044141e38bcefc93f1718657). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.384209] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-635ff849-e798-424e-a157-bc849b4ab96b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.399967] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Created folder: Project (ea8a6296044141e38bcefc93f1718657) in parent group-v201431. [ 717.400215] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Creating folder: Instances. Parent ref: group-v201470. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 717.400473] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7108abbb-7265-4386-a23b-cee2b3bc3ad5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.414358] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Created folder: Instances in parent group-v201470. [ 717.414630] env[62000]: DEBUG oslo.service.loopingcall [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 717.415070] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 717.415304] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dbc9064b-b65e-4fbc-8ac9-1c6d6d0fe438 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.434182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a377345b-b503-4a12-8f61-be98713ce46e tempest-ListServerFiltersTestJSON-1034416581 tempest-ListServerFiltersTestJSON-1034416581-project-member] Lock "1419e5a8-a78f-44f1-bfc4-00fcf176bf35" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.789s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.439220] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 717.439220] env[62000]: value = "task-881934" [ 717.439220] env[62000]: _type = "Task" [ 717.439220] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.452784] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881934, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.475193] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881925, 'name': Rename_Task, 'duration_secs': 1.731793} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.475891] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 717.476206] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0156073a-1ba2-45d5-9976-2a9974ff01b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.481721] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881930, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.483508] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 717.483508] env[62000]: value = "task-881935" [ 717.483508] env[62000]: _type = "Task" [ 717.483508] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.491825] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881935, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.785439] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881929, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.793370] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881931, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059622} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.793881] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 717.794766] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16cded41-4d73-4f9e-96cb-077f57a6703e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.817682] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 717.817682] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27216d83-7c5a-49ed-95db-ce1b9565d6a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.838164] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 717.838164] env[62000]: value = "task-881936" [ 717.838164] env[62000]: _type = "Task" [ 717.838164] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.846526] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881936, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 717.923959] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.953933] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881934, 'name': CreateVM_Task, 'duration_secs': 0.494856} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.954138] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 717.954931] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.955097] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.955460] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 717.955696] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47fe1e03-dae5-4b14-8249-7964df9c8ce7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.961717] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 717.961717] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d21e7-af41-92ff-c0ce-19479d99adbb" [ 717.961717] env[62000]: _type = "Task" [ 717.961717] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 717.980644] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d21e7-af41-92ff-c0ce-19479d99adbb, 'name': SearchDatastore_Task, 'duration_secs': 0.009814} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.983261] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.983517] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 717.983752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.984031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.984099] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 717.986653] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25689894-b9b8-4515-94e1-37c18608e85d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.998298] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881930, 'name': PowerOffVM_Task, 'duration_secs': 0.573583} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 717.998298] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 717.998298] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 717.998298] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c07d852f-9ed8-4b7b-8e72-fda1d0df0c45 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.005278] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881935, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.009271] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 718.009504] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 718.010585] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a61b426-db2d-4b4b-b460-fe77ffe5640a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.017870] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 718.017870] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5268e2c9-ad4a-cb88-2d81-7c3f9cf2b9e2" [ 718.017870] env[62000]: _type = "Task" [ 718.017870] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.029016] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5268e2c9-ad4a-cb88-2d81-7c3f9cf2b9e2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.124465] env[62000]: DEBUG nova.network.neutron [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.184973] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 718.185116] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 718.185835] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleting the datastore file [datastore2] 1d9b67b5-c0cf-41ff-a838-79caf8789609 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 718.185835] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65709d02-e03a-45cc-b0f6-dc52963af0a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.195248] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 718.195248] env[62000]: value = "task-881938" [ 718.195248] env[62000]: _type = "Task" [ 718.195248] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.203337] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881938, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.267055] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72d7ec2-9c9e-47cb-972d-f52b2fef0e43 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.277465] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f288794f-f20e-46c4-b227-3839bfe01339 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.285113] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881929, 'name': ReconfigVM_Task, 'duration_secs': 1.045531} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.309773] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Reconfigured VM instance instance-00000026 to attach disk [datastore1] 415cc4a5-7610-4678-971d-cd00a0e8b54d/415cc4a5-7610-4678-971d-cd00a0e8b54d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.310991] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-77cabe4a-a4ab-474c-9d65-a9fd154613db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.313308] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fc1ddd9-ade8-4e43-bf23-531d8b9adce9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.324137] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39efdba6-a87e-4522-a8db-21b1bf84ffc9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.327798] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 718.327798] env[62000]: value = "task-881939" [ 718.327798] env[62000]: _type = "Task" [ 718.327798] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.341020] env[62000]: DEBUG nova.compute.provider_tree [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 718.345871] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881939, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.353017] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881936, 'name': ReconfigVM_Task, 'duration_secs': 0.306409} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.353017] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 718.353017] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b2d89bcb-5115-4d3c-bde3-91b83a3e3f6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.362135] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 718.362135] env[62000]: value = "task-881940" [ 718.362135] env[62000]: _type = "Task" [ 718.362135] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.371034] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881940, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.496269] env[62000]: DEBUG oslo_vmware.api [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-881935, 'name': PowerOnVM_Task, 'duration_secs': 0.563538} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.496469] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 718.496767] env[62000]: INFO nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Took 12.79 seconds to spawn the instance on the hypervisor. [ 718.497077] env[62000]: DEBUG nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 718.498037] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-114953d0-3ade-40ce-aa36-89aa12e67765 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.529437] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5268e2c9-ad4a-cb88-2d81-7c3f9cf2b9e2, 'name': SearchDatastore_Task, 'duration_secs': 0.010221} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.530251] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23d698fd-bf6c-451b-a2e4-8accbc13c53e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.535947] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 718.535947] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278c5ef-e6b0-430a-b674-5352ac6e2f21" [ 718.535947] env[62000]: _type = "Task" [ 718.535947] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.543389] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278c5ef-e6b0-430a-b674-5352ac6e2f21, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.629442] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.629805] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Instance network_info: |[{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 718.630319] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3d:56:f3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '365ac5b1-6d83-4dfe-887f-60574d7f6124', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1d7778e6-8a84-447d-a241-95ea69ab68ba', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 718.637851] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Creating folder: Project (c051bb06d4c04417902263736e94a8b0). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.638157] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-237dbbeb-85f4-4f08-8c0c-b1833d255110 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.649617] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Created folder: Project (c051bb06d4c04417902263736e94a8b0) in parent group-v201431. [ 718.649822] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Creating folder: Instances. Parent ref: group-v201473. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 718.650072] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-224a968b-ecab-461c-824a-c48f373a9c1e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.659346] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Created folder: Instances in parent group-v201473. [ 718.659589] env[62000]: DEBUG oslo.service.loopingcall [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.659821] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 718.660022] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-929ad37d-8690-46eb-9a97-723e5c23083a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.679105] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 718.679105] env[62000]: value = "task-881943" [ 718.679105] env[62000]: _type = "Task" [ 718.679105] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.686375] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881943, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.697571] env[62000]: DEBUG nova.compute.manager [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 718.698440] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32d9af8c-4f90-4bb4-a20b-794db2c9e88a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.709428] env[62000]: DEBUG oslo_vmware.api [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-881938, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169238} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.711044] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 718.711250] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 718.711435] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 718.711614] env[62000]: INFO nova.compute.manager [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Took 1.76 seconds to destroy the instance on the hypervisor. [ 718.712042] env[62000]: DEBUG oslo.service.loopingcall [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.714331] env[62000]: DEBUG nova.compute.manager [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.714396] env[62000]: DEBUG nova.network.neutron [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.837352] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881939, 'name': Rename_Task, 'duration_secs': 0.25067} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.845508] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 718.845508] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-32031a80-d60a-42f0-9a63-fb8dab15e43f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.845508] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 718.845508] env[62000]: value = "task-881944" [ 718.845508] env[62000]: _type = "Task" [ 718.845508] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.856654] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881944, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.870854] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881940, 'name': Rename_Task, 'duration_secs': 0.1474} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 718.871159] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 718.871417] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b70556fc-27e1-4e00-8743-8f28bf62705f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.876663] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 718.876663] env[62000]: value = "task-881945" [ 718.876663] env[62000]: _type = "Task" [ 718.876663] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 718.885035] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881945, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 718.885877] env[62000]: DEBUG nova.scheduler.client.report [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 59 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 718.886127] env[62000]: DEBUG nova.compute.provider_tree [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 59 to 60 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 718.886313] env[62000]: DEBUG nova.compute.provider_tree [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 719.024780] env[62000]: INFO nova.compute.manager [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Took 33.50 seconds to build instance. [ 719.024780] env[62000]: DEBUG nova.compute.manager [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.024780] env[62000]: DEBUG nova.compute.manager [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing instance network info cache due to event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 719.024780] env[62000]: DEBUG oslo_concurrency.lockutils [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.024780] env[62000]: DEBUG oslo_concurrency.lockutils [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.024780] env[62000]: DEBUG nova.network.neutron [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 719.051404] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278c5ef-e6b0-430a-b674-5352ac6e2f21, 'name': SearchDatastore_Task, 'duration_secs': 0.011602} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.051771] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.052218] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 72688992-333f-459d-9d05-f7c728961a6d/72688992-333f-459d-9d05-f7c728961a6d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 719.052503] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-31bc6ac9-4169-4132-a351-7beeff5336f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.061226] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 719.061226] env[62000]: value = "task-881946" [ 719.061226] env[62000]: _type = "Task" [ 719.061226] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.073098] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881946, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.191109] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881943, 'name': CreateVM_Task, 'duration_secs': 0.433463} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.191311] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 719.192070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.192275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.192563] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 719.193226] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e9418652-1db4-409e-9477-e981f1b54047 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.199094] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 719.199094] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5256fd31-e8a6-d1ba-b7bc-377d44a092f9" [ 719.199094] env[62000]: _type = "Task" [ 719.199094] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.208878] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5256fd31-e8a6-d1ba-b7bc-377d44a092f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.218130] env[62000]: INFO nova.compute.manager [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] instance snapshotting [ 719.221823] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c37e097a-ee2b-455a-be09-cc7f95287a86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.249912] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f801b59-6bb8-45ba-be92-b463f7180487 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.357762] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881944, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.393044] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.486s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.394025] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 719.398135] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881945, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.399520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.453s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.402957] env[62000]: INFO nova.compute.claims [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.526567] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e79f0653-74b6-49c1-aa38-7652df7bd07a tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.110s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.572097] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881946, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.712851] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5256fd31-e8a6-d1ba-b7bc-377d44a092f9, 'name': SearchDatastore_Task, 'duration_secs': 0.00955} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.713347] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.713710] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 719.716151] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.717394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.717394] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 719.717394] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e048777a-e9ad-4c42-b65b-6b766adf2510 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.736360] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 719.736360] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 719.736862] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-489d181b-1a06-4137-ab3b-371c09fd46ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.744893] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 719.744893] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bbd8e7-9680-ea55-3973-90ab0f0223f8" [ 719.744893] env[62000]: _type = "Task" [ 719.744893] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.755532] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bbd8e7-9680-ea55-3973-90ab0f0223f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.766957] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 719.767265] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-14346f1e-6a79-49da-9919-44ae5fc8a8ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.776683] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 719.776683] env[62000]: value = "task-881947" [ 719.776683] env[62000]: _type = "Task" [ 719.776683] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 719.786907] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 719.857234] env[62000]: DEBUG oslo_vmware.api [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881944, 'name': PowerOnVM_Task, 'duration_secs': 0.519079} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.857659] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 719.857750] env[62000]: INFO nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Took 11.68 seconds to spawn the instance on the hypervisor. [ 719.857892] env[62000]: DEBUG nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.858709] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1db256c5-c7ca-4543-979e-2c79782e6817 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.891215] env[62000]: DEBUG oslo_vmware.api [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-881945, 'name': PowerOnVM_Task, 'duration_secs': 0.538143} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 719.891339] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 719.891474] env[62000]: INFO nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 9.17 seconds to spawn the instance on the hypervisor. [ 719.891649] env[62000]: DEBUG nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 719.893075] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9739c13-ed53-40b5-b9ca-d78e79d6a493 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.900993] env[62000]: DEBUG nova.compute.utils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 719.908661] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 719.908838] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 719.972062] env[62000]: DEBUG nova.compute.manager [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Received event network-changed-c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 719.972345] env[62000]: DEBUG nova.compute.manager [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Refreshing instance network info cache due to event network-changed-c0059cab-c384-49ca-b0fa-1e46163e688c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 719.972641] env[62000]: DEBUG oslo_concurrency.lockutils [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] Acquiring lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.972768] env[62000]: DEBUG oslo_concurrency.lockutils [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] Acquired lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.972995] env[62000]: DEBUG nova.network.neutron [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Refreshing network info cache for port c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 720.028243] env[62000]: DEBUG nova.policy [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6712f85a79744b71a2924a66fdcc5bf4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cbd14025e565405a93d619f2ac64c28f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 720.031567] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 720.076909] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881946, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643838} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.077029] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 72688992-333f-459d-9d05-f7c728961a6d/72688992-333f-459d-9d05-f7c728961a6d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 720.077348] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 720.077692] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8082e6d1-6081-4def-8b6f-c771329601ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.087908] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 720.087908] env[62000]: value = "task-881948" [ 720.087908] env[62000]: _type = "Task" [ 720.087908] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.099448] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881948, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.155368] env[62000]: DEBUG nova.network.neutron [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updated VIF entry in instance network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 720.156634] env[62000]: DEBUG nova.network.neutron [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.258282] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bbd8e7-9680-ea55-3973-90ab0f0223f8, 'name': SearchDatastore_Task, 'duration_secs': 0.055994} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.259161] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-deaa9ab5-b60d-4f19-b260-9c0d58868730 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.265319] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 720.265319] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64106-f537-1839-3164-fb277bdb1d8c" [ 720.265319] env[62000]: _type = "Task" [ 720.265319] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.280382] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64106-f537-1839-3164-fb277bdb1d8c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.290949] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.323038] env[62000]: DEBUG nova.network.neutron [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.379618] env[62000]: INFO nova.compute.manager [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Took 32.99 seconds to build instance. [ 720.404586] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 720.422571] env[62000]: INFO nova.compute.manager [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 26.24 seconds to build instance. [ 720.456560] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Successfully created port: d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 720.559263] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.598011] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881948, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075971} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.600657] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 720.602193] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74efc485-94d5-4cfa-a288-57fb10393733 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.625947] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Reconfiguring VM instance instance-00000028 to attach disk [datastore1] 72688992-333f-459d-9d05-f7c728961a6d/72688992-333f-459d-9d05-f7c728961a6d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 720.630681] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-23e36e81-833c-4a9d-a02a-c3be85012cab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.655571] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 720.655571] env[62000]: value = "task-881949" [ 720.655571] env[62000]: _type = "Task" [ 720.655571] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.660248] env[62000]: DEBUG oslo_concurrency.lockutils [req-681ba07e-2741-4827-a6e4-01ee81164b79 req-376a2838-5a4c-4e76-9e04-c6f66812ef7e service nova] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.670115] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881949, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.787590] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64106-f537-1839-3164-fb277bdb1d8c, 'name': SearchDatastore_Task, 'duration_secs': 0.031532} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 720.790454] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.790807] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/64ff4a7b-ec89-48cd-8fb6-124e0726d6f0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 720.791430] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b076f57f-8473-4fa8-b66b-e3a49cf9d339 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.796527] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.803476] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 720.803476] env[62000]: value = "task-881950" [ 720.803476] env[62000]: _type = "Task" [ 720.803476] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 720.813342] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881950, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 720.819533] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1da81a-e9f6-4773-8431-aa444a6684bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.829375] env[62000]: INFO nova.compute.manager [-] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Took 2.11 seconds to deallocate network for instance. [ 720.832131] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249fecd3-4292-4958-a649-2a9850a9d432 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.872147] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16ed1f66-e135-46cf-8841-25ab64397b91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.880867] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c8bc03e-8f55-44d6-9fcc-91a637c2ac2f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.887118] env[62000]: DEBUG oslo_concurrency.lockutils [None req-768179e2-f300-463e-b45e-696d80448d36 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.489s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.899520] env[62000]: DEBUG nova.compute.provider_tree [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.912524] env[62000]: INFO nova.virt.block_device [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Booting with volume 5684c204-cdf0-4896-8023-0152376f3fb3 at /dev/sda [ 720.927885] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f657e52a-cb6b-44c6-a31a-b987108e7a9f tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 124.692s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.980021] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-43530f43-3451-4ced-a281-5eb46ce9f629 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.990059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b82c1cd-4eb5-480f-964a-3cef26e01ca9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.027677] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7355aaf8-baeb-4ce2-91de-447290134419 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.036326] env[62000]: DEBUG nova.network.neutron [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updated VIF entry in instance network info cache for port c0059cab-c384-49ca-b0fa-1e46163e688c. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 721.036885] env[62000]: DEBUG nova.network.neutron [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updating instance_info_cache with network_info: [{"id": "c0059cab-c384-49ca-b0fa-1e46163e688c", "address": "fa:16:3e:89:bf:85", "network": {"id": "c4fade47-cb22-4277-86be-849d1fceb2b2", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1349651934-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.165", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "76de44563cdd4a3883d5153555c2e48e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f847601f-7479-48eb-842f-41f94eea8537", "external-id": "nsx-vlan-transportzone-35", "segmentation_id": 35, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc0059cab-c3", "ovs_interfaceid": "c0059cab-c384-49ca-b0fa-1e46163e688c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.040749] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8b1b0a6-585e-41f4-80fb-6cccc84fd67b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.076699] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-293285c0-8492-4594-952c-6d4b26320d74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.086301] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fcb4a2-7a64-4dac-b45b-86bcc7e34404 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.104068] env[62000]: DEBUG nova.virt.block_device [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updating existing volume attachment record: f47f1521-35f0-4a5a-abb7-b4ad04c0abfc {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 721.166928] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881949, 'name': ReconfigVM_Task, 'duration_secs': 0.366034} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.167256] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Reconfigured VM instance instance-00000028 to attach disk [datastore1] 72688992-333f-459d-9d05-f7c728961a6d/72688992-333f-459d-9d05-f7c728961a6d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 721.168023] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8a760039-e21c-4719-beb3-6ae997fee459 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.174622] env[62000]: DEBUG nova.compute.manager [None req-b6251fac-3a99-4f49-9341-b46283b50e49 tempest-ServerDiagnosticsV248Test-229552800 tempest-ServerDiagnosticsV248Test-229552800-project-admin] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 721.175357] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c19e6ce-59fa-4f23-aa9a-2e71f4b17fdc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.179820] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 721.179820] env[62000]: value = "task-881951" [ 721.179820] env[62000]: _type = "Task" [ 721.179820] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.186710] env[62000]: INFO nova.compute.manager [None req-b6251fac-3a99-4f49-9341-b46283b50e49 tempest-ServerDiagnosticsV248Test-229552800 tempest-ServerDiagnosticsV248Test-229552800-project-admin] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Retrieving diagnostics [ 721.187986] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f696bf2b-752e-459f-bdc3-da8ee64c46bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.195591] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881951, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.235229] env[62000]: DEBUG nova.compute.manager [req-d255cdcb-561a-42a8-a9af-1cf4d135942d req-7154cfe0-29d7-4e5d-8fd6-892881125d91 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-deleted-10672d07-d10e-48f7-b383-87348134adb9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 721.235597] env[62000]: DEBUG nova.compute.manager [req-d255cdcb-561a-42a8-a9af-1cf4d135942d req-7154cfe0-29d7-4e5d-8fd6-892881125d91 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-deleted-3d6933ed-ca06-4521-a288-8c4c4cacefd1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 721.235597] env[62000]: DEBUG nova.compute.manager [req-d255cdcb-561a-42a8-a9af-1cf4d135942d req-7154cfe0-29d7-4e5d-8fd6-892881125d91 service nova] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Received event network-vif-deleted-556236e8-eb70-4dc1-8fbd-5090a7df2605 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 721.291660] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.315023] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881950, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.342854] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.390081] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.404938] env[62000]: DEBUG nova.scheduler.client.report [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.433830] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.553807] env[62000]: DEBUG oslo_concurrency.lockutils [req-966e1190-17c0-48b7-acee-57b601d8a72a req-8c89278f-c5ed-486e-9528-c4d355c42eea service nova] Releasing lock "refresh_cache-49f6c198-13b7-4c07-81d8-c010e7b0598e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.703324] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881951, 'name': Rename_Task, 'duration_secs': 0.208188} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.703324] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 721.703324] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7f610b45-ce29-430a-a9cc-275df4e7d35f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.711906] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 721.711906] env[62000]: value = "task-881952" [ 721.711906] env[62000]: _type = "Task" [ 721.711906] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.720716] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881952, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.791866] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.815614] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881950, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.570368} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 721.815717] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/64ff4a7b-ec89-48cd-8fb6-124e0726d6f0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 721.815883] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 721.816155] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5f054681-b48d-4478-9584-939b2cabcca3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 721.825184] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 721.825184] env[62000]: value = "task-881953" [ 721.825184] env[62000]: _type = "Task" [ 721.825184] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 721.835661] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881953, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 721.910796] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.911455] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.915849] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.900s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.916277] env[62000]: DEBUG nova.objects.instance [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lazy-loading 'resources' on Instance uuid 762e429f-f690-43c4-95eb-877caf1cdad7 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 721.921271] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.969071] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.229511] env[62000]: DEBUG oslo_vmware.api [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881952, 'name': PowerOnVM_Task, 'duration_secs': 0.509267} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.230171] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 722.230545] env[62000]: INFO nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Took 9.00 seconds to spawn the instance on the hypervisor. [ 722.230638] env[62000]: DEBUG nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 722.231808] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56eede71-647a-486c-9764-feec949e994d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.294705] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881947, 'name': CreateSnapshot_Task, 'duration_secs': 2.205698} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.294913] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 722.295947] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23882f91-df1e-438b-8ed4-22b496a4f06f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.337792] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881953, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073315} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 722.338569] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 722.340033] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60115b5-e1a9-40ff-b87c-596f394dd59c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.373486] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/64ff4a7b-ec89-48cd-8fb6-124e0726d6f0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 722.374236] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-904587a4-ea84-40eb-bd26-b754eef79ace {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.396815] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 722.396815] env[62000]: value = "task-881954" [ 722.396815] env[62000]: _type = "Task" [ 722.396815] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.406824] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.419504] env[62000]: DEBUG nova.compute.utils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 722.424205] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 722.424492] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 722.493412] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.493688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.493905] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.494109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.494317] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.496545] env[62000]: INFO nova.compute.manager [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Terminating instance [ 722.500768] env[62000]: DEBUG nova.compute.manager [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 722.500972] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 722.502085] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82613622-60fb-49a5-b440-c9619967d62f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.510757] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 722.511818] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52be38a3-f29e-4b7c-94bb-bdda15a3f81f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.519659] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 722.519659] env[62000]: value = "task-881955" [ 722.519659] env[62000]: _type = "Task" [ 722.519659] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.534120] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.690459] env[62000]: DEBUG nova.policy [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aec03d6033bb47cd8542e15c627b25e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd2a32e7597d640c0a17818e965dbf4d3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.707076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.708374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.708374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.708541] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 722.708832] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.712917] env[62000]: INFO nova.compute.manager [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Terminating instance [ 722.718570] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 722.718741] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquired lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 722.718915] env[62000]: DEBUG nova.network.neutron [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 722.756959] env[62000]: INFO nova.compute.manager [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Took 26.16 seconds to build instance. [ 722.817887] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 722.818699] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0e3bdaa9-9acf-417c-9d66-12b461385985 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.827594] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e57ebba2-1a34-4c24-ad96-691e4286678d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.836665] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acfd329a-e6c6-46b6-ab6b-06fe5076dbcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.842811] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 722.842811] env[62000]: value = "task-881956" [ 722.842811] env[62000]: _type = "Task" [ 722.842811] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 722.872406] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-048446e5-faa3-4fc6-9088-47d561464cf1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.879057] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881956, 'name': CloneVM_Task} progress is 11%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.885351] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7825e5b-efc9-4fa6-a905-7dffc4f6d690 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.905076] env[62000]: DEBUG nova.compute.provider_tree [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 722.917953] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881954, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 722.926182] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 723.034610] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881955, 'name': PowerOffVM_Task, 'duration_secs': 0.429411} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.034948] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 723.035032] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 723.035309] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-838bba14-84a2-45a9-95dc-5316293ccd4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.060188] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Successfully updated port: d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 723.123360] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 723.123360] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 723.123697] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Deleting the datastore file [datastore1] 415cc4a5-7610-4678-971d-cd00a0e8b54d {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 723.123891] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62bfb145-75a5-4524-8514-e02282927949 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.134285] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for the task: (returnval){ [ 723.134285] env[62000]: value = "task-881958" [ 723.134285] env[62000]: _type = "Task" [ 723.134285] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.146675] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.248746] env[62000]: DEBUG nova.network.neutron [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.254609] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.256043] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.256255] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.256351] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.256493] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.256660] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.257061] env[62000]: DEBUG nova.virt.hardware [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.257688] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ec400ea-b43c-4b07-9a11-f7b3885c1ab5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.260519] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cda1cba2-fafd-4600-96e8-1e4c222a71c4 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.040s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.270920] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4176a410-d857-4329-954d-293afb07e680 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.347951] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Successfully created port: 892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 723.354036] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881956, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.382407] env[62000]: DEBUG nova.network.neutron [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.414464] env[62000]: DEBUG nova.scheduler.client.report [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.418746] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881954, 'name': ReconfigVM_Task, 'duration_secs': 0.709386} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.419261] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/64ff4a7b-ec89-48cd-8fb6-124e0726d6f0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 723.420173] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5148f9ae-4774-48b7-b222-6d0842a3d3f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.429305] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 723.429305] env[62000]: value = "task-881959" [ 723.429305] env[62000]: _type = "Task" [ 723.429305] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.445937] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881959, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.562226] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Received event network-vif-plugged-d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.562669] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Acquiring lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.563092] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.563480] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.563828] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] No waiting events found dispatching network-vif-plugged-d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 723.564221] env[62000]: WARNING nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Received unexpected event network-vif-plugged-d89856ae-9afc-460b-bc77-5a3635aa59f7 for instance with vm_state building and task_state spawning. [ 723.564567] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Received event network-changed-a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 723.564918] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Refreshing instance network info cache due to event network-changed-a40f58be-07f1-43ce-afd9-e0ea0c78b634. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 723.565307] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.565644] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.566008] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Refreshing network info cache for port a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 723.568232] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.568510] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquired lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.568803] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.647522] env[62000]: DEBUG oslo_vmware.api [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Task: {'id': task-881958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.373146} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 723.648220] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 723.648379] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 723.648563] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 723.648740] env[62000]: INFO nova.compute.manager [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Took 1.15 seconds to destroy the instance on the hypervisor. [ 723.649640] env[62000]: DEBUG oslo.service.loopingcall [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 723.649640] env[62000]: DEBUG nova.compute.manager [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 723.649640] env[62000]: DEBUG nova.network.neutron [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.763508] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.859250] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881956, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.886481] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Releasing lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.886481] env[62000]: DEBUG nova.compute.manager [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 723.886972] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 723.887876] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf176e78-37ac-40e1-94e8-91fe1527c38e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.896455] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 723.896807] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14ee52d0-c630-4212-a499-d26b075de015 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.908169] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 723.908169] env[62000]: value = "task-881960" [ 723.908169] env[62000]: _type = "Task" [ 723.908169] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 723.918602] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.920530] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.005s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.922868] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.879s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.924493] env[62000]: INFO nova.compute.claims [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.941812] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.943976] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881959, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 723.948926] env[62000]: INFO nova.scheduler.client.report [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Deleted allocations for instance 762e429f-f690-43c4-95eb-877caf1cdad7 [ 723.970870] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.970870] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.970870] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.972802] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.972802] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.972802] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.972802] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.973138] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.973378] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.973778] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.973862] env[62000]: DEBUG nova.virt.hardware [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.975733] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1d958f-ffe2-484a-9de4-fb8b0c586eb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.987493] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9ccf31b-9ca2-4eea-81fc-5bb0cc037657 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.148451] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.287820] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.357927] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881956, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.378808] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.379099] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.418733] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881960, 'name': PowerOffVM_Task, 'duration_secs': 0.274555} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.419483] env[62000]: DEBUG nova.network.neutron [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updating instance_info_cache with network_info: [{"id": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "address": "fa:16:3e:9d:b1:1c", "network": {"id": "3f54f1ab-b381-433a-898d-9680054a6797", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1006300929-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbd14025e565405a93d619f2ac64c28f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd89856ae-9a", "ovs_interfaceid": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.420608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 724.420802] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 724.421067] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e4ea0fe8-7ecc-4818-be06-96fa31c3b46c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.454973] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 724.455350] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 724.455644] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleting the datastore file [datastore2] 5e80a5c8-030c-4ad8-90c4-26136fa39d71 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 724.457033] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1c941906-942f-4675-87ac-4d83767cfaac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.469791] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881959, 'name': Rename_Task, 'duration_secs': 0.793784} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.470429] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cff23a3b-ded8-4201-b067-55cc8807654f tempest-InstanceActionsV221TestJSON-945743975 tempest-InstanceActionsV221TestJSON-945743975-project-member] Lock "762e429f-f690-43c4-95eb-877caf1cdad7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.921s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 724.475029] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 724.475029] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60aa2f0a-bcb3-4b81-84e2-7588a135f17b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.478115] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for the task: (returnval){ [ 724.478115] env[62000]: value = "task-881962" [ 724.478115] env[62000]: _type = "Task" [ 724.478115] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.490117] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 724.490117] env[62000]: value = "task-881963" [ 724.490117] env[62000]: _type = "Task" [ 724.490117] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.500746] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.507057] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881963, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 724.507173] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updated VIF entry in instance network info cache for port a40f58be-07f1-43ce-afd9-e0ea0c78b634. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 724.507445] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.561224] env[62000]: DEBUG nova.network.neutron [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.824334] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8709d5a-c1df-4db1-8e88-1644255e23be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.832535] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-197790cd-16b6-4d28-8729-a0881f35cc02 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.865607] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9925cc25-54c0-4713-9ffb-8ac0e1ac0342 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.873728] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881956, 'name': CloneVM_Task, 'duration_secs': 1.919218} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.876214] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Created linked-clone VM from snapshot [ 724.877095] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd249a5d-873a-4f07-831c-a9bd9252c61e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.882201] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42a14ea-aaf1-4bb8-8c2d-83ab90aae8c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.890523] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Uploading image d6794f74-f464-4ec1-b5cc-c7b585923b03 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 724.900931] env[62000]: DEBUG nova.compute.provider_tree [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.924456] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Releasing lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.924785] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Instance network_info: |[{"id": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "address": "fa:16:3e:9d:b1:1c", "network": {"id": "3f54f1ab-b381-433a-898d-9680054a6797", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1006300929-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbd14025e565405a93d619f2ac64c28f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd89856ae-9a", "ovs_interfaceid": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 724.925213] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9d:b1:1c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3836fb52-19c6-4e10-a0ca-f0bca73dc887', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd89856ae-9afc-460b-bc77-5a3635aa59f7', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 724.932735] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Creating folder: Project (cbd14025e565405a93d619f2ac64c28f). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.935335] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5def30e0-9be6-4463-a2fe-46bc2476722d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.945501] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 724.945501] env[62000]: value = "vm-201477" [ 724.945501] env[62000]: _type = "VirtualMachine" [ 724.945501] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 724.945811] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-e576623c-64d2-4756-8cf9-75051a793ffe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.950774] env[62000]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 724.950774] env[62000]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62000) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 724.951112] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Folder already exists: Project (cbd14025e565405a93d619f2ac64c28f). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 724.951320] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Creating folder: Instances. Parent ref: group-v201437. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 724.951972] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9765f380-b0b8-4cc9-ab70-958b2022b4a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.955458] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lease: (returnval){ [ 724.955458] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5232d8dc-e5a2-7e1b-86d1-98697e68a4fd" [ 724.955458] env[62000]: _type = "HttpNfcLease" [ 724.955458] env[62000]: } obtained for exporting VM: (result){ [ 724.955458] env[62000]: value = "vm-201477" [ 724.955458] env[62000]: _type = "VirtualMachine" [ 724.955458] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 724.955806] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the lease: (returnval){ [ 724.955806] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5232d8dc-e5a2-7e1b-86d1-98697e68a4fd" [ 724.955806] env[62000]: _type = "HttpNfcLease" [ 724.955806] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 724.962599] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 724.962599] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5232d8dc-e5a2-7e1b-86d1-98697e68a4fd" [ 724.962599] env[62000]: _type = "HttpNfcLease" [ 724.962599] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 724.963840] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Created folder: Instances in parent group-v201437. [ 724.964063] env[62000]: DEBUG oslo.service.loopingcall [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.964293] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 724.964511] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-909dbb9e-dd0c-46fc-b2fc-3c8762486ae9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.990150] env[62000]: DEBUG oslo_vmware.api [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Task: {'id': task-881962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.204097} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 724.991413] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 724.991684] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 724.991807] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 724.992060] env[62000]: INFO nova.compute.manager [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Took 1.11 seconds to destroy the instance on the hypervisor. [ 724.992329] env[62000]: DEBUG oslo.service.loopingcall [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 724.992531] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 724.992531] env[62000]: value = "task-881967" [ 724.992531] env[62000]: _type = "Task" [ 724.992531] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 724.992806] env[62000]: DEBUG nova.compute.manager [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 724.992907] env[62000]: DEBUG nova.network.neutron [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 725.006356] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881963, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.009875] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881967, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.012599] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.012819] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Received event network-changed-d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.013009] env[62000]: DEBUG nova.compute.manager [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Refreshing instance network info cache due to event network-changed-d89856ae-9afc-460b-bc77-5a3635aa59f7. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 725.013250] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Acquiring lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.013421] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Acquired lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.013582] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Refreshing network info cache for port d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.036020] env[62000]: DEBUG nova.network.neutron [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.064834] env[62000]: INFO nova.compute.manager [-] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Took 1.42 seconds to deallocate network for instance. [ 725.268678] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "72688992-333f-459d-9d05-f7c728961a6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.269052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.269315] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "72688992-333f-459d-9d05-f7c728961a6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.269538] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.269746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.276679] env[62000]: INFO nova.compute.manager [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Terminating instance [ 725.282828] env[62000]: DEBUG nova.compute.manager [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.292529] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 725.292529] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b61ec4a8-4176-4e1c-8ff5-58bd4112c6f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.297535] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 725.297937] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a87cd62-ce08-422b-a80a-f75e73402a6b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.310015] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 725.310015] env[62000]: value = "task-881968" [ 725.310015] env[62000]: _type = "Task" [ 725.310015] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.315981] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881968, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.403871] env[62000]: DEBUG nova.scheduler.client.report [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.412995] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Successfully updated port: 892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 725.469020] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 725.469020] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5232d8dc-e5a2-7e1b-86d1-98697e68a4fd" [ 725.469020] env[62000]: _type = "HttpNfcLease" [ 725.469020] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 725.469020] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 725.469020] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5232d8dc-e5a2-7e1b-86d1-98697e68a4fd" [ 725.469020] env[62000]: _type = "HttpNfcLease" [ 725.469020] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 725.469020] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57ee8519-b744-4630-836f-2129352bc1d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.476688] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 725.477039] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 725.542739] env[62000]: DEBUG nova.network.neutron [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.559378] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881967, 'name': CreateVM_Task, 'duration_secs': 0.366464} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.560356] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 725.561120] env[62000]: DEBUG oslo_vmware.api [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881963, 'name': PowerOnVM_Task, 'duration_secs': 0.633656} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.561765] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'f47f1521-35f0-4a5a-abb7-b4ad04c0abfc', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201440', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'name': 'volume-5684c204-cdf0-4896-8023-0152376f3fb3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '334fd514-c7b5-402f-935e-1d95f9b5dbc8', 'attached_at': '', 'detached_at': '', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'serial': '5684c204-cdf0-4896-8023-0152376f3fb3'}, 'mount_device': '/dev/sda', 'device_type': None, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62000) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 725.562117] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Root volume attach. Driver type: vmdk {{(pid=62000) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 725.562491] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 725.562739] env[62000]: INFO nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Took 9.81 seconds to spawn the instance on the hypervisor. [ 725.562951] env[62000]: DEBUG nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 725.563862] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87aeaf1d-db4f-4c3b-bb81-c66785020032 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.569421] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5466c0-26ac-42eb-b849-b95d466fcf5b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.572718] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.580069] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e0f89d-d826-4e7c-abd6-96e1ac23178d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.586388] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b001d9fb-1cb4-461a-a382-ae720757a703 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.598406] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-5ac8e580-cf5a-4392-b099-c6234b5debb4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.603482] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 725.603482] env[62000]: value = "task-881969" [ 725.603482] env[62000]: _type = "Task" [ 725.603482] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.611851] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.617580] env[62000]: DEBUG nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Received event network-vif-deleted-f077ad17-a015-4c19-aa71-f1f595d55c65 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.617833] env[62000]: DEBUG nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Received event network-vif-plugged-892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.621022] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Acquiring lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 725.621022] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.621022] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.621022] env[62000]: DEBUG nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] No waiting events found dispatching network-vif-plugged-892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 725.621022] env[62000]: WARNING nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Received unexpected event network-vif-plugged-892944d2-5b6b-4a93-b975-f9e1f8104b1e for instance with vm_state building and task_state spawning. [ 725.621022] env[62000]: DEBUG nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Received event network-changed-892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 725.621022] env[62000]: DEBUG nova.compute.manager [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Refreshing instance network info cache due to event network-changed-892944d2-5b6b-4a93-b975-f9e1f8104b1e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 725.621022] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Acquiring lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.621022] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Acquired lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.621022] env[62000]: DEBUG nova.network.neutron [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Refreshing network info cache for port 892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 725.653199] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-7bc64864-1904-4d93-aeb4-8eb2f0bab15d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.817037] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881968, 'name': PowerOffVM_Task, 'duration_secs': 0.19406} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 725.817333] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 725.817520] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 725.817772] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca164941-9796-4b60-8939-6d0563028ffd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.832706] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updated VIF entry in instance network info cache for port d89856ae-9afc-460b-bc77-5a3635aa59f7. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 725.833111] env[62000]: DEBUG nova.network.neutron [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updating instance_info_cache with network_info: [{"id": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "address": "fa:16:3e:9d:b1:1c", "network": {"id": "3f54f1ab-b381-433a-898d-9680054a6797", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1006300929-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbd14025e565405a93d619f2ac64c28f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd89856ae-9a", "ovs_interfaceid": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.883620] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 725.884022] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 725.884310] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Deleting the datastore file [datastore1] 72688992-333f-459d-9d05-f7c728961a6d {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 725.884610] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3c37d683-ac26-4d55-a769-60a66e3965cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.891611] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for the task: (returnval){ [ 725.891611] env[62000]: value = "task-881971" [ 725.891611] env[62000]: _type = "Task" [ 725.891611] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 725.901933] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881971, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 725.908771] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.986s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.910076] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 725.912579] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.185s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.914298] env[62000]: INFO nova.compute.claims [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.917879] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.054044] env[62000]: INFO nova.compute.manager [-] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Took 1.06 seconds to deallocate network for instance. [ 726.094862] env[62000]: INFO nova.compute.manager [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Took 27.51 seconds to build instance. [ 726.119620] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 35%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.174535] env[62000]: DEBUG nova.network.neutron [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.312891] env[62000]: DEBUG nova.network.neutron [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.335964] env[62000]: DEBUG oslo_concurrency.lockutils [req-becd2439-2adc-4a16-8121-4e0a53f3e1a8 req-f1066e4b-43ec-4ff1-975f-c542e5bcaefa service nova] Releasing lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.407178] env[62000]: DEBUG oslo_vmware.api [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Task: {'id': task-881971, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.240812} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 726.407393] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 726.407492] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 726.408320] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 726.409026] env[62000]: INFO nova.compute.manager [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Took 1.13 seconds to destroy the instance on the hypervisor. [ 726.409357] env[62000]: DEBUG oslo.service.loopingcall [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 726.409584] env[62000]: DEBUG nova.compute.manager [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 726.409688] env[62000]: DEBUG nova.network.neutron [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.421841] env[62000]: DEBUG nova.compute.utils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.427511] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.430778] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 726.543442] env[62000]: DEBUG nova.policy [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4770851cc3b4ad8aaa6866a18ee1359', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41e10f7d4f964f2795dc629721802880', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 726.562505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.597998] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9033f09a-b7ed-4c48-8391-1a85cca6d9a6 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.391s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.617639] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 49%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 726.822192] env[62000]: DEBUG oslo_concurrency.lockutils [req-16ccb154-8ab7-4084-b247-f626ab97259c req-aad6d2e3-ef6d-4c12-a10d-48ed31cb8f18 service nova] Releasing lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.822750] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquired lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.823054] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 726.942092] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 726.994719] env[62000]: INFO nova.compute.manager [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Rescuing [ 726.995063] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.995217] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.995383] env[62000]: DEBUG nova.network.neutron [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.100794] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 727.118632] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 63%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.266781] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Successfully created port: 1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.367634] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c881c8-1330-4056-aac7-1bc108f0218b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.380359] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ea1aeb-f017-407b-8203-625887ac5aa5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.385364] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 727.422906] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24498e03-fc75-4b50-84c8-255ca88882de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.437504] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba3b9a37-106c-43bd-9d27-a9e9fc708d9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.468074] env[62000]: DEBUG nova.compute.provider_tree [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.626692] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 76%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 727.630853] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.674210] env[62000]: DEBUG nova.network.neutron [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.775282] env[62000]: DEBUG nova.compute.manager [req-e4b6bf6c-b918-4826-a129-4ccb2cbd5103 req-6b0290c0-f3b5-4f5d-ae8a-109966384b86 service nova] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Received event network-vif-deleted-b63c8316-148a-47a4-b17a-d394a10a072c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 727.819065] env[62000]: DEBUG nova.network.neutron [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Updating instance_info_cache with network_info: [{"id": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "address": "fa:16:3e:47:3b:52", "network": {"id": "941616e2-425d-4828-847e-c1f3f4b49d34", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-381389191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2a32e7597d640c0a17818e965dbf4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892944d2-5b", "ovs_interfaceid": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.970534] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 727.973279] env[62000]: DEBUG nova.scheduler.client.report [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 728.003184] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.003480] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.003641] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.003851] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.003973] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.004375] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.004508] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.004705] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.004921] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.005159] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.005360] env[62000]: DEBUG nova.virt.hardware [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.006637] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-290fccfe-f786-4eb3-88f3-757cd77685dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.018114] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e285f347-dd0a-4c6b-8120-f7a87eee3328 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.040239] env[62000]: DEBUG nova.network.neutron [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.123869] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.177034] env[62000]: INFO nova.compute.manager [-] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Took 1.77 seconds to deallocate network for instance. [ 728.322425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Releasing lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.322771] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance network_info: |[{"id": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "address": "fa:16:3e:47:3b:52", "network": {"id": "941616e2-425d-4828-847e-c1f3f4b49d34", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-381389191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2a32e7597d640c0a17818e965dbf4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892944d2-5b", "ovs_interfaceid": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 728.323242] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:47:3b:52', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'ba4f6497-e2b4-43b5-9819-6927865ae974', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '892944d2-5b6b-4a93-b975-f9e1f8104b1e', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 728.330809] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Creating folder: Project (d2a32e7597d640c0a17818e965dbf4d3). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 728.331165] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d816f8a-d2de-4775-b8d8-f92db5df6378 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.342752] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Created folder: Project (d2a32e7597d640c0a17818e965dbf4d3) in parent group-v201431. [ 728.342965] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Creating folder: Instances. Parent ref: group-v201480. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 728.343263] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-573d5347-dd35-4c77-a0fd-29ff398dc606 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.351940] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Created folder: Instances in parent group-v201480. [ 728.352228] env[62000]: DEBUG oslo.service.loopingcall [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.352440] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 728.352653] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cb11950-4494-492e-9526-9eb425b20236 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.372150] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 728.372150] env[62000]: value = "task-881974" [ 728.372150] env[62000]: _type = "Task" [ 728.372150] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.385989] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881974, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.482030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.568s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.482030] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.484955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.579s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.487150] env[62000]: INFO nova.compute.claims [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.543834] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.621187] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.685358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.881985] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881974, 'name': CreateVM_Task, 'duration_secs': 0.503251} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 728.883156] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 728.883920] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.884116] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.884486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 728.884765] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e44cf4d4-7603-48b9-96a5-79d0c5ffd9b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.889482] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 728.889482] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5227251d-b278-3612-ba11-a214e25a1bbc" [ 728.889482] env[62000]: _type = "Task" [ 728.889482] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 728.897593] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5227251d-b278-3612-ba11-a214e25a1bbc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 728.993627] env[62000]: DEBUG nova.compute.utils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.998616] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.998771] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 729.078597] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 729.078873] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-31da8601-cf08-4074-bc53-6b95d173bc4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.087455] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 729.087455] env[62000]: value = "task-881975" [ 729.087455] env[62000]: _type = "Task" [ 729.087455] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.096794] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881975, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.120337] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task} progress is 98%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.148621] env[62000]: DEBUG nova.policy [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0fabd6e8940941ada9540441560393a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad12e169457d491097ab311e05d8e30a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.402995] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5227251d-b278-3612-ba11-a214e25a1bbc, 'name': SearchDatastore_Task, 'duration_secs': 0.052121} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.403333] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.403574] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.403819] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.403971] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.404234] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.404444] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdde1744-c0b1-435b-bd41-c6eae22f2ae2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.413563] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.414203] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 729.415047] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff256619-e0e0-4d95-9709-8b84643caa9e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.421074] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 729.421074] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52226afa-342a-1ecb-8b71-17319490c759" [ 729.421074] env[62000]: _type = "Task" [ 729.421074] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.433385] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52226afa-342a-1ecb-8b71-17319490c759, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.499288] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.599725] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881975, 'name': PowerOffVM_Task, 'duration_secs': 0.208127} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.600172] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 729.600978] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f49d209-227d-445a-b303-5974d70728a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.628050] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8f0ed7-cde7-41fb-ac64-82260689a0ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.638573] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881969, 'name': RelocateVM_Task, 'duration_secs': 3.802686} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.641125] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 729.641125] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201440', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'name': 'volume-5684c204-cdf0-4896-8023-0152376f3fb3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '334fd514-c7b5-402f-935e-1d95f9b5dbc8', 'attached_at': '', 'detached_at': '', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'serial': '5684c204-cdf0-4896-8023-0152376f3fb3'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 729.645967] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f0c230-9282-403d-b4fa-e81ff8ec898f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.667172] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c517f85-1cb5-4374-8abc-f6a4d73eeffb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.689609] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] volume-5684c204-cdf0-4896-8023-0152376f3fb3/volume-5684c204-cdf0-4896-8023-0152376f3fb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 729.694057] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 729.694349] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fd471ab-766d-4c81-bc1e-e082f68b6cbb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.708637] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96b54899-73a3-44f9-9451-7263386a62fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.716438] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 729.716438] env[62000]: value = "task-881977" [ 729.716438] env[62000]: _type = "Task" [ 729.716438] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.717686] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 729.717686] env[62000]: value = "task-881976" [ 729.717686] env[62000]: _type = "Task" [ 729.717686] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.730933] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881977, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.734575] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 729.734786] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 729.735110] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.735283] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.735498] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 729.735842] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fbe88594-be4d-4c77-8ded-a03f28d62c7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.746914] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 729.747179] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 729.748157] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f4770f6-3fff-479f-a6c0-5602366159fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.757623] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 729.757623] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b26e4e-4f28-6a94-a728-8a8239c41273" [ 729.757623] env[62000]: _type = "Task" [ 729.757623] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.768446] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b26e4e-4f28-6a94-a728-8a8239c41273, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 729.822128] env[62000]: DEBUG nova.compute.manager [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Received event network-vif-plugged-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 729.822476] env[62000]: DEBUG oslo_concurrency.lockutils [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] Acquiring lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.822743] env[62000]: DEBUG oslo_concurrency.lockutils [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.823155] env[62000]: DEBUG oslo_concurrency.lockutils [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.823259] env[62000]: DEBUG nova.compute.manager [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] No waiting events found dispatching network-vif-plugged-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 729.823387] env[62000]: WARNING nova.compute.manager [req-4782e47c-50d7-4a97-a973-b446ff067927 req-3afc9c53-69ff-482d-9075-68762a0f58eb service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Received unexpected event network-vif-plugged-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 for instance with vm_state building and task_state spawning. [ 729.845852] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Successfully updated port: 1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 729.928865] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Successfully created port: bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.937181] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52226afa-342a-1ecb-8b71-17319490c759, 'name': SearchDatastore_Task, 'duration_secs': 0.014049} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.938027] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-307f6cc7-5285-4166-a769-299c95e907bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.943356] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2459916-71c2-447d-abf2-516a2bd70ec7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.947332] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 729.947332] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52416901-df96-f4e3-95c2-300211012e98" [ 729.947332] env[62000]: _type = "Task" [ 729.947332] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 729.955441] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df434e7e-13a4-47af-8770-ef5c7aea388d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.962558] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52416901-df96-f4e3-95c2-300211012e98, 'name': SearchDatastore_Task, 'duration_secs': 0.01093} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 729.963239] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.963546] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 210af329-4cdb-4c3f-9e82-e72a2ea79421/210af329-4cdb-4c3f-9e82-e72a2ea79421.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 729.963824] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89f08a03-de55-4c22-8ca1-fcf548e47c71 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.996069] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f9f228d-2992-4984-82b0-427915d01594 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.999257] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 729.999257] env[62000]: value = "task-881978" [ 729.999257] env[62000]: _type = "Task" [ 729.999257] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.006292] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52a5c5e2-8038-4fbb-bd7b-ae0f9856871d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.016358] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.026800] env[62000]: DEBUG nova.compute.provider_tree [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.234677] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881977, 'name': ReconfigVM_Task, 'duration_secs': 0.251948} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.235061] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Reconfigured VM instance instance-0000002a to attach disk [datastore2] volume-5684c204-cdf0-4896-8023-0152376f3fb3/volume-5684c204-cdf0-4896-8023-0152376f3fb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 730.241947] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-827c1b51-7c19-439e-9df2-9f5c272d7582 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.260398] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 730.260398] env[62000]: value = "task-881979" [ 730.260398] env[62000]: _type = "Task" [ 730.260398] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.276721] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b26e4e-4f28-6a94-a728-8a8239c41273, 'name': SearchDatastore_Task, 'duration_secs': 0.011405} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.282405] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881979, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.282800] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4f3da11-0056-4588-9174-8fbf7ddb7ec4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.290665] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 730.290665] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d02734-9712-862f-2b67-08b577ee87f2" [ 730.290665] env[62000]: _type = "Task" [ 730.290665] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.301139] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d02734-9712-862f-2b67-08b577ee87f2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.347561] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.349649] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.349649] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 730.512165] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881978, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.521202] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.532019] env[62000]: DEBUG nova.scheduler.client.report [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.552300] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.553239] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.553239] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.553239] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.553239] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.553530] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.553530] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.553683] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.553847] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.554022] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.554243] env[62000]: DEBUG nova.virt.hardware [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.555761] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c679d21f-4b50-40c6-89f8-39287c130b9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.565355] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2190899e-0c55-4cad-a39b-e97f477c1a65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.777421] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881979, 'name': ReconfigVM_Task, 'duration_secs': 0.196112} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.777421] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201440', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'name': 'volume-5684c204-cdf0-4896-8023-0152376f3fb3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '334fd514-c7b5-402f-935e-1d95f9b5dbc8', 'attached_at': '', 'detached_at': '', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'serial': '5684c204-cdf0-4896-8023-0152376f3fb3'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 730.777941] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6a5a5f16-c695-4c92-9fa2-d8a9a6721f87 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.787690] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 730.787690] env[62000]: value = "task-881980" [ 730.787690] env[62000]: _type = "Task" [ 730.787690] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.804927] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d02734-9712-862f-2b67-08b577ee87f2, 'name': SearchDatastore_Task, 'duration_secs': 0.017339} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 730.809014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 730.809434] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72-rescue.vmdk. {{(pid=62000) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 730.809771] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881980, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.810155] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efd4c472-ee3b-4d28-8ca8-a17278c54054 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.817044] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 730.817044] env[62000]: value = "task-881981" [ 730.817044] env[62000]: _type = "Task" [ 730.817044] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 730.827313] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881981, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 730.896119] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.014311] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881978, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.572381} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.014311] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 210af329-4cdb-4c3f-9e82-e72a2ea79421/210af329-4cdb-4c3f-9e82-e72a2ea79421.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 731.014571] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 731.015389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4a62df04-29ba-424e-8ff8-95c1c5ff2bc1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.022806] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 731.022806] env[62000]: value = "task-881982" [ 731.022806] env[62000]: _type = "Task" [ 731.022806] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.039019] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881982, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.039019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.552s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.039019] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 731.040101] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.481s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.041582] env[62000]: INFO nova.compute.claims [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.147296] env[62000]: DEBUG nova.network.neutron [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Updating instance_info_cache with network_info: [{"id": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "address": "fa:16:3e:e4:58:db", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fd8ccbc-54", "ovs_interfaceid": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.299752] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881980, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.328748] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881981, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.534527] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881982, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.124386} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.535394] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 731.536388] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb5a63c-5c2a-461a-9d4b-7d1df52d23fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.552423] env[62000]: DEBUG nova.compute.utils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.564340] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Reconfiguring VM instance instance-0000002b to attach disk [datastore2] 210af329-4cdb-4c3f-9e82-e72a2ea79421/210af329-4cdb-4c3f-9e82-e72a2ea79421.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.565920] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.566321] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 731.572222] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2aff9377-9f31-4524-a171-1ee1098a38a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.600610] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 731.600610] env[62000]: value = "task-881983" [ 731.600610] env[62000]: _type = "Task" [ 731.600610] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.611218] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881983, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.649694] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.649694] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance network_info: |[{"id": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "address": "fa:16:3e:e4:58:db", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fd8ccbc-54", "ovs_interfaceid": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 731.651739] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:58:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 731.663179] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating folder: Project (41e10f7d4f964f2795dc629721802880). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 731.663179] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-764df487-4595-4cb9-8222-4b85d79c10c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.667851] env[62000]: DEBUG nova.policy [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '074bf610faa641eaa1d05467c31719f2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '087cacaac4d64fcc85cc3fcad3ee9c2e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.677924] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created folder: Project (41e10f7d4f964f2795dc629721802880) in parent group-v201431. [ 731.678464] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating folder: Instances. Parent ref: group-v201483. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 731.678464] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dc7e1fbf-f7b7-48c1-81fc-4fdfe5d06ea4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.689366] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created folder: Instances in parent group-v201483. [ 731.689627] env[62000]: DEBUG oslo.service.loopingcall [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 731.689846] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 731.690140] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b9e83d0-5100-46ef-8dd7-700fc3bf7103 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.713807] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 731.713807] env[62000]: value = "task-881986" [ 731.713807] env[62000]: _type = "Task" [ 731.713807] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.723889] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881986, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.797925] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881980, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 731.829487] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881981, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.540963} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 731.829820] env[62000]: INFO nova.virt.vmwareapi.ds_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72-rescue.vmdk. [ 731.830699] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41654fe5-ab6a-487b-893f-194d31e9e5bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.863168] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Reconfiguring VM instance instance-00000029 to attach disk [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72-rescue.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 731.863168] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15846c1a-dab2-4ebb-9035-3fc826393306 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.886238] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 731.886238] env[62000]: value = "task-881987" [ 731.886238] env[62000]: _type = "Task" [ 731.886238] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 731.896563] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881987, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.066728] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 732.123150] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881983, 'name': ReconfigVM_Task, 'duration_secs': 0.46791} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.123150] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Reconfigured VM instance instance-0000002b to attach disk [datastore2] 210af329-4cdb-4c3f-9e82-e72a2ea79421/210af329-4cdb-4c3f-9e82-e72a2ea79421.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.123150] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-936401c1-4a08-4364-9583-1153b78f64e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.124630] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Successfully updated port: bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 732.127623] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 732.127623] env[62000]: value = "task-881988" [ 732.127623] env[62000]: _type = "Task" [ 732.127623] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.147141] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881988, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.165307] env[62000]: DEBUG nova.compute.manager [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Received event network-changed-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 732.165307] env[62000]: DEBUG nova.compute.manager [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Refreshing instance network info cache due to event network-changed-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 732.165307] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Acquiring lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.165307] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Acquired lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.165307] env[62000]: DEBUG nova.network.neutron [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Refreshing network info cache for port 1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 732.224257] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881986, 'name': CreateVM_Task, 'duration_secs': 0.430403} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.224418] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 732.225246] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.225335] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.225679] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 732.228469] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec709f35-5c54-49f6-b1cd-ca9ff51d5f38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.234759] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 732.234759] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52616aa6-eb7b-25f3-eedd-caa17ec687dd" [ 732.234759] env[62000]: _type = "Task" [ 732.234759] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.243755] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52616aa6-eb7b-25f3-eedd-caa17ec687dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.283108] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Successfully created port: e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.301137] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881980, 'name': Rename_Task, 'duration_secs': 1.112811} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.301137] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.301497] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-547f78fd-8e95-475b-a252-a29ad322617e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.309140] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 732.309140] env[62000]: value = "task-881989" [ 732.309140] env[62000]: _type = "Task" [ 732.309140] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.319574] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.402567] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881987, 'name': ReconfigVM_Task, 'duration_secs': 0.319668} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.402930] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Reconfigured VM instance instance-00000029 to attach disk [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72-rescue.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 732.403970] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4969d67b-df4f-4da3-ad6e-f82ffc14b6a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.438130] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9875f7d-3268-4c1e-ba5f-94899c0871f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.457011] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 732.457011] env[62000]: value = "task-881990" [ 732.457011] env[62000]: _type = "Task" [ 732.457011] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.467627] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881990, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.520503] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ae8b31e-60aa-4d28-abe7-d297a1f0be11 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.528716] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217ac9ac-0f0e-42d0-b120-34c574889d79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.563452] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9874bd43-50bd-4d00-aeca-cd6eb59f88c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.572614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c94ff0-b47f-473e-95fb-1f680e1ea09c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.589393] env[62000]: DEBUG nova.compute.provider_tree [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.628648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.628836] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.629014] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.639760] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881988, 'name': Rename_Task, 'duration_secs': 0.214968} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.640113] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.640382] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e2c88dd4-c360-4867-8686-81b686eba98e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.646462] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 732.646462] env[62000]: value = "task-881991" [ 732.646462] env[62000]: _type = "Task" [ 732.646462] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.656445] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881991, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.746168] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52616aa6-eb7b-25f3-eedd-caa17ec687dd, 'name': SearchDatastore_Task, 'duration_secs': 0.011207} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.746589] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.746761] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 732.747243] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.747465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.747675] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 732.748050] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ae2f01a-cc0c-4e48-937c-4c2f64d266d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.758897] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 732.759198] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 732.760208] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d459d5c0-d92a-4d6a-9091-990ff250b34b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.767946] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 732.767946] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c44fc8-4646-cab4-2644-8553ed57a8a5" [ 732.767946] env[62000]: _type = "Task" [ 732.767946] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.781167] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c44fc8-4646-cab4-2644-8553ed57a8a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 732.820758] env[62000]: DEBUG oslo_vmware.api [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-881989, 'name': PowerOnVM_Task, 'duration_secs': 0.489335} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.821059] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 732.821270] env[62000]: INFO nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Took 9.57 seconds to spawn the instance on the hypervisor. [ 732.821456] env[62000]: DEBUG nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 732.822875] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd24d939-e66f-441c-a583-125ad0f725ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.965492] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881990, 'name': ReconfigVM_Task, 'duration_secs': 0.176564} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 732.966030] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 732.966402] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-595dd2f5-51bc-4b04-a8b8-1df0b4b6a142 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.974169] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 732.974169] env[62000]: value = "task-881992" [ 732.974169] env[62000]: _type = "Task" [ 732.974169] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 732.987906] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881992, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.014038] env[62000]: DEBUG nova.network.neutron [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Updated VIF entry in instance network info cache for port 1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 733.014897] env[62000]: DEBUG nova.network.neutron [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Updating instance_info_cache with network_info: [{"id": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "address": "fa:16:3e:e4:58:db", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1fd8ccbc-54", "ovs_interfaceid": "1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.084465] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 733.092732] env[62000]: DEBUG nova.scheduler.client.report [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 733.118168] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 733.118509] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 733.118747] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 733.118984] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 733.119192] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 733.119681] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 733.120098] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 733.120351] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 733.120570] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 733.120779] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 733.120996] env[62000]: DEBUG nova.virt.hardware [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 733.122196] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd16d92f-1933-4fef-bc38-d8b07fcfbd41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.136518] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8393d164-003d-4100-a089-9dd48588b39c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.170656] env[62000]: DEBUG oslo_vmware.api [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-881991, 'name': PowerOnVM_Task, 'duration_secs': 0.50345} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.170947] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 733.171355] env[62000]: INFO nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Took 9.23 seconds to spawn the instance on the hypervisor. [ 733.171488] env[62000]: DEBUG nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 733.172380] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be2fc86-62c1-48f6-b97f-f8bfef27f35e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.184654] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.280044] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c44fc8-4646-cab4-2644-8553ed57a8a5, 'name': SearchDatastore_Task, 'duration_secs': 0.020423} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.284111] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca5fb79f-d784-4949-9098-6e0f48f5d036 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.291187] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 733.291187] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e31e22-d051-8747-b848-99a76f29d570" [ 733.291187] env[62000]: _type = "Task" [ 733.291187] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.300669] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e31e22-d051-8747-b848-99a76f29d570, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.342372] env[62000]: INFO nova.compute.manager [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Took 32.66 seconds to build instance. [ 733.382021] env[62000]: DEBUG nova.network.neutron [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.486291] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881992, 'name': PowerOnVM_Task} progress is 96%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.519378] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Releasing lock "refresh_cache-596af5ab-1791-42ce-93d2-3e4f0a47dfa2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.519378] env[62000]: DEBUG nova.compute.manager [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received event network-vif-plugged-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 733.519378] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Acquiring lock "4e617bb5-84e4-4505-99e3-61289826f511-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.519378] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Lock "4e617bb5-84e4-4505-99e3-61289826f511-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.519378] env[62000]: DEBUG oslo_concurrency.lockutils [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] Lock "4e617bb5-84e4-4505-99e3-61289826f511-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.519378] env[62000]: DEBUG nova.compute.manager [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] No waiting events found dispatching network-vif-plugged-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 733.519378] env[62000]: WARNING nova.compute.manager [req-a688edf5-3ffe-4da1-a39d-ffaa13ee85fa req-4173b9d9-1884-4c6c-8198-3fdd44ab08d9 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received unexpected event network-vif-plugged-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 for instance with vm_state building and task_state spawning. [ 733.606082] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.566s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.606626] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.614019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.267s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.614019] env[62000]: DEBUG nova.objects.instance [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lazy-loading 'resources' on Instance uuid 1d9b67b5-c0cf-41ff-a838-79caf8789609 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 733.693348] env[62000]: INFO nova.compute.manager [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Took 28.76 seconds to build instance. [ 733.804637] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e31e22-d051-8747-b848-99a76f29d570, 'name': SearchDatastore_Task, 'duration_secs': 0.020937} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.804938] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.805218] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 733.805509] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c220b667-10bb-464f-955e-398de25b6b2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.812362] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 733.812362] env[62000]: value = "task-881993" [ 733.812362] env[62000]: _type = "Task" [ 733.812362] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.822023] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881993, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.845499] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3a8a9dd9-274c-4866-b3ee-48ba8cd55615 tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.707s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.885155] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.886023] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Instance network_info: |[{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 733.886394] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:71:46:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e1049e8-c06b-4c93-a9e1-2cbb530f3f95', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcf8ac20-4efb-4ad5-a14a-9548cfb59944', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 733.895018] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Creating folder: Project (ad12e169457d491097ab311e05d8e30a). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.895705] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-51ae303a-2731-4847-9313-a250b9214b3a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.906882] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Created folder: Project (ad12e169457d491097ab311e05d8e30a) in parent group-v201431. [ 733.906882] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Creating folder: Instances. Parent ref: group-v201486. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 733.906882] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e65e5e3f-5a19-4115-9c2d-5e460fea6ed0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.916615] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Created folder: Instances in parent group-v201486. [ 733.917336] env[62000]: DEBUG oslo.service.loopingcall [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 733.917739] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 733.919015] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-384208a4-9622-41a0-8721-3d33253f8928 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.958444] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 733.958444] env[62000]: value = "task-881996" [ 733.958444] env[62000]: _type = "Task" [ 733.958444] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 733.972992] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881996, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 733.989373] env[62000]: DEBUG oslo_vmware.api [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-881992, 'name': PowerOnVM_Task, 'duration_secs': 0.592406} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 733.989882] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 733.996190] env[62000]: DEBUG nova.compute.manager [None req-f135adcb-3ad1-4989-aa4c-704fae996505 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 733.997534] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f93e3fd-6826-4e56-8ec2-4de4b146f232 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.116974] env[62000]: DEBUG nova.compute.utils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 734.119397] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 734.195905] env[62000]: DEBUG oslo_concurrency.lockutils [None req-442f175c-4249-4820-9f56-6244c004b27a tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.498s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.323921] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881993, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.350614] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.421178] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Successfully updated port: e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 734.474758] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881996, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.493861] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-234278b1-8908-4eb8-9222-c44974955d00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.501630] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2a9b69-bbc5-4890-9f35-6efbb1fc0445 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.540038] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31fccad7-b7f1-45bb-9aa3-0b25a62d1966 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.545417] env[62000]: DEBUG nova.compute.manager [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 734.545627] env[62000]: DEBUG nova.compute.manager [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing instance network info cache due to event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 734.545861] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Acquiring lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.546015] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Acquired lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.548061] env[62000]: DEBUG nova.network.neutron [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 734.555383] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76b6e4e-f302-44ce-abe0-cad51da42b7f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.570498] env[62000]: DEBUG nova.compute.provider_tree [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.588917] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "2b0c7354-1622-4318-ba09-6e3214eed4db" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.589093] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 734.621319] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.703635] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.713124] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 734.713822] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db199db1-3ca2-4623-91d8-6f3411249276 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.725774] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 734.725774] env[62000]: ERROR oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk due to incomplete transfer. [ 734.726029] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-33dc3196-8b58-4ed1-bb7d-085d188b3fe6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.732931] env[62000]: DEBUG oslo_vmware.rw_handles [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/521de677-a9a5-28fe-19ac-05f8cc46f3fd/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 734.733360] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Uploaded image d6794f74-f464-4ec1-b5cc-c7b585923b03 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 734.735407] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 734.735663] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-20714741-54e5-4555-9fcf-4a600a5cd4b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.742291] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 734.742291] env[62000]: value = "task-881997" [ 734.742291] env[62000]: _type = "Task" [ 734.742291] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.754433] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881997, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.823743] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881993, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567852} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 734.823967] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 734.824221] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 734.824486] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-df2f6c17-0f0b-4c3f-8109-bb9703b28e72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.832068] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 734.832068] env[62000]: value = "task-881998" [ 734.832068] env[62000]: _type = "Task" [ 734.832068] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 734.839356] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881998, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 734.873134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.923496] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 734.923633] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquired lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 734.923827] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 734.972732] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881996, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.074042] env[62000]: DEBUG nova.scheduler.client.report [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.230746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.256274] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881997, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.343774] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881998, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062061} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.344077] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 735.345162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f5ca955-ae93-44bd-a927-3f06307e6d03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.369636] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Reconfiguring VM instance instance-0000002c to attach disk [datastore2] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 735.375083] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9ea37b2-9b61-403e-9349-e8c5b5041e1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.395403] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 735.395403] env[62000]: value = "task-881999" [ 735.395403] env[62000]: _type = "Task" [ 735.395403] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.403497] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881999, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.470979] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881996, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.502054] env[62000]: DEBUG nova.compute.manager [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Received event network-changed-d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 735.502353] env[62000]: DEBUG nova.compute.manager [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Refreshing instance network info cache due to event network-changed-d89856ae-9afc-460b-bc77-5a3635aa59f7. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 735.502676] env[62000]: DEBUG oslo_concurrency.lockutils [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] Acquiring lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.502896] env[62000]: DEBUG oslo_concurrency.lockutils [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] Acquired lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.503111] env[62000]: DEBUG nova.network.neutron [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Refreshing network info cache for port d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 735.512536] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 735.580258] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.970s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.582975] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.662s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.584376] env[62000]: INFO nova.compute.claims [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 735.609603] env[62000]: INFO nova.scheduler.client.report [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleted allocations for instance 1d9b67b5-c0cf-41ff-a838-79caf8789609 [ 735.630599] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.663147] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.663479] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.663647] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.663843] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.663986] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.664210] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.664478] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.664679] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.664827] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.664999] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.665264] env[62000]: DEBUG nova.virt.hardware [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.666387] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9b748a-3b9a-4ea8-824e-d0bb95f7849e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.670623] env[62000]: DEBUG nova.network.neutron [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updated VIF entry in instance network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 735.671022] env[62000]: DEBUG nova.network.neutron [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.679491] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d1f947e-731c-4f79-b12b-196cf6895854 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.703793] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 735.711149] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Creating folder: Project (113e60ce7bc54e2cb179543b9ca664b0). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 735.711943] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-78b54d78-0967-459d-be2a-b3d095d524c6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.725019] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Created folder: Project (113e60ce7bc54e2cb179543b9ca664b0) in parent group-v201431. [ 735.725242] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Creating folder: Instances. Parent ref: group-v201489. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 735.725483] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a90040eb-1a80-4bbc-b5d6-6e2b2c8b57ed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.737635] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Created folder: Instances in parent group-v201489. [ 735.738222] env[62000]: DEBUG oslo.service.loopingcall [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 735.738947] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 735.739280] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d593cac9-90fe-4afb-bf1f-7bf5494cbce4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.767767] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881997, 'name': Destroy_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.769232] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 735.769232] env[62000]: value = "task-882002" [ 735.769232] env[62000]: _type = "Task" [ 735.769232] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.778793] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882002, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.906556] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-881999, 'name': ReconfigVM_Task, 'duration_secs': 0.451521} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.906908] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Reconfigured VM instance instance-0000002c to attach disk [datastore2] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 735.907585] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1e572a07-9e9f-44fb-829a-9fc3e4d6c861 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.914206] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 735.914206] env[62000]: value = "task-882003" [ 735.914206] env[62000]: _type = "Task" [ 735.914206] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.922571] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882003, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 735.928994] env[62000]: DEBUG nova.network.neutron [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Updating instance_info_cache with network_info: [{"id": "e01b6797-8122-495f-b322-2212f250481b", "address": "fa:16:3e:fa:bb:6e", "network": {"id": "a4de1643-552e-424b-ba9b-17ff06d7270d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1878559773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "087cacaac4d64fcc85cc3fcad3ee9c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape01b6797-81", "ovs_interfaceid": "e01b6797-8122-495f-b322-2212f250481b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 735.972278] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-881996, 'name': CreateVM_Task, 'duration_secs': 1.651019} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 735.972495] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 735.973231] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.973468] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.973855] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 735.974329] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7ac12c49-cf0c-4845-baca-3233a77feee3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.979208] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 735.979208] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c4bb5e-5de8-d3c2-7dfc-48a88ef6cefa" [ 735.979208] env[62000]: _type = "Task" [ 735.979208] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 735.987257] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c4bb5e-5de8-d3c2-7dfc-48a88ef6cefa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.117423] env[62000]: DEBUG oslo_concurrency.lockutils [None req-da4682c6-48dc-4c51-8d23-57a242b2aa13 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "1d9b67b5-c0cf-41ff-a838-79caf8789609" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.174s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.174638] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Releasing lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.175045] env[62000]: DEBUG nova.compute.manager [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Received event network-vif-plugged-e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.175154] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Acquiring lock "af49714d-8e50-4159-96a5-cf8f70580471-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 736.175359] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Lock "af49714d-8e50-4159-96a5-cf8f70580471-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.175520] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] Lock "af49714d-8e50-4159-96a5-cf8f70580471-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.175689] env[62000]: DEBUG nova.compute.manager [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] No waiting events found dispatching network-vif-plugged-e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 736.175854] env[62000]: WARNING nova.compute.manager [req-9f5a5863-1187-406d-9d0b-467279fe6989 req-6f859625-9349-49dd-96a3-25dc78520095 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Received unexpected event network-vif-plugged-e01b6797-8122-495f-b322-2212f250481b for instance with vm_state building and task_state spawning. [ 736.270225] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-881997, 'name': Destroy_Task, 'duration_secs': 1.120005} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.270225] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Destroyed the VM [ 736.270492] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 736.276232] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9bce22f7-2d9b-42f7-870e-c8fbb7e21868 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.283236] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882002, 'name': CreateVM_Task, 'duration_secs': 0.299777} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.285101] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 736.285101] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 736.285101] env[62000]: value = "task-882004" [ 736.285101] env[62000]: _type = "Task" [ 736.285101] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.285429] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.294792] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882004, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.369685] env[62000]: DEBUG nova.network.neutron [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updated VIF entry in instance network info cache for port d89856ae-9afc-460b-bc77-5a3635aa59f7. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 736.370365] env[62000]: DEBUG nova.network.neutron [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updating instance_info_cache with network_info: [{"id": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "address": "fa:16:3e:9d:b1:1c", "network": {"id": "3f54f1ab-b381-433a-898d-9680054a6797", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-1006300929-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.183", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cbd14025e565405a93d619f2ac64c28f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3836fb52-19c6-4e10-a0ca-f0bca73dc887", "external-id": "nsx-vlan-transportzone-964", "segmentation_id": 964, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd89856ae-9a", "ovs_interfaceid": "d89856ae-9afc-460b-bc77-5a3635aa59f7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.425080] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882003, 'name': Rename_Task, 'duration_secs': 0.148168} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.425512] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 736.425964] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cbeeb43b-9f82-42da-9742-57257019d966 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.431391] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Releasing lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.431779] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Instance network_info: |[{"id": "e01b6797-8122-495f-b322-2212f250481b", "address": "fa:16:3e:fa:bb:6e", "network": {"id": "a4de1643-552e-424b-ba9b-17ff06d7270d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1878559773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "087cacaac4d64fcc85cc3fcad3ee9c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape01b6797-81", "ovs_interfaceid": "e01b6797-8122-495f-b322-2212f250481b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 736.433414] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:bb:6e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2b6c45fd-e930-495a-9cb7-df84eda443b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e01b6797-8122-495f-b322-2212f250481b', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.441063] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Creating folder: Project (087cacaac4d64fcc85cc3fcad3ee9c2e). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 736.441492] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 736.441492] env[62000]: value = "task-882005" [ 736.441492] env[62000]: _type = "Task" [ 736.441492] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.441715] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-98074a30-38dc-4e62-a414-f4e5e400ee3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.452487] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882005, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.454538] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Created folder: Project (087cacaac4d64fcc85cc3fcad3ee9c2e) in parent group-v201431. [ 736.454755] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Creating folder: Instances. Parent ref: group-v201492. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 736.455016] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa3e67a2-580b-4ccc-8cf5-d6abe9042813 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.464280] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Created folder: Instances in parent group-v201492. [ 736.464564] env[62000]: DEBUG oslo.service.loopingcall [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.464791] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 736.465028] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-447c243c-12f9-438c-9378-e44680d187fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.491167] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c4bb5e-5de8-d3c2-7dfc-48a88ef6cefa, 'name': SearchDatastore_Task, 'duration_secs': 0.009772} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.492603] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.493063] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 736.493254] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.493427] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.493636] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 736.493921] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.493921] env[62000]: value = "task-882008" [ 736.493921] env[62000]: _type = "Task" [ 736.493921] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.494170] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.494535] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 736.494765] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-99259b33-bf62-4688-b326-fc9468c2183b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.496796] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-722b36eb-b713-4fd9-9d96-0bc91116fb66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.505994] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 736.505994] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ec07b-2e3b-b3b8-91d1-0c59b5e26772" [ 736.505994] env[62000]: _type = "Task" [ 736.505994] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.510440] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882008, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.514684] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 736.514684] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 736.515478] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd444167-a5af-46f0-b6bc-32a97af92987 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.523852] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ec07b-2e3b-b3b8-91d1-0c59b5e26772, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.525678] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 736.525678] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c5490-5cf8-880a-8006-a1f6f1d031ea" [ 736.525678] env[62000]: _type = "Task" [ 736.525678] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.537934] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c5490-5cf8-880a-8006-a1f6f1d031ea, 'name': SearchDatastore_Task, 'duration_secs': 0.010791} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 736.540922] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfa3c59f-a570-4d27-ae4e-cd3492527a93 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.545746] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 736.545746] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5267a32a-10ba-f8f8-6f28-29c2c379138e" [ 736.545746] env[62000]: _type = "Task" [ 736.545746] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.554494] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5267a32a-10ba-f8f8-6f28-29c2c379138e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.799031] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882004, 'name': RemoveSnapshot_Task} progress is 84%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.865321] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Received event network-changed-e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 736.865592] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Refreshing instance network info cache due to event network-changed-e01b6797-8122-495f-b322-2212f250481b. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 736.865727] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquiring lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.865875] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquired lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.866070] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Refreshing network info cache for port e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 736.876992] env[62000]: DEBUG oslo_concurrency.lockutils [req-7d066a5c-2f6a-4f62-91d0-288cc0ace5cc req-45ae3141-e9dd-4ca0-b875-8ec6747f3201 service nova] Releasing lock "refresh_cache-334fd514-c7b5-402f-935e-1d95f9b5dbc8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.958695] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882005, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 736.979561] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58ba42b0-80ad-4230-8fa2-705a36fe46d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.988106] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fefc6932-b3cb-4513-b5f8-908b8c28aa11 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.036619] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22ce9e0a-a62b-4d28-9606-31d1b7f6fd6e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.054033] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab1f1b25-d5d0-4bfd-83c8-28f7969f93b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.055667] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882008, 'name': CreateVM_Task, 'duration_secs': 0.39329} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.056013] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527ec07b-2e3b-b3b8-91d1-0c59b5e26772, 'name': SearchDatastore_Task, 'duration_secs': 0.018203} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.059200] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 737.059539] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.059774] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.060224] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.061391] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.061555] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.061932] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.062500] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2801b8a-4907-4dac-b36d-cdaf4b9127f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.077243] env[62000]: DEBUG nova.compute.provider_tree [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.078708] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5267a32a-10ba-f8f8-6f28-29c2c379138e, 'name': SearchDatastore_Task, 'duration_secs': 0.010104} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.079489] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.079789] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4e617bb5-84e4-4505-99e3-61289826f511/4e617bb5-84e4-4505-99e3-61289826f511.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 737.080140] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.080384] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 737.080627] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2d8a70f0-86d1-49e6-8b76-f3028ad95da5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.084300] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9d7bf616-91cc-41d9-a487-f76aa6e3ba81 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.086263] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 737.086263] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522655e6-69c3-7c49-5eaf-1be57ac8775d" [ 737.086263] env[62000]: _type = "Task" [ 737.086263] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.097581] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522655e6-69c3-7c49-5eaf-1be57ac8775d, 'name': SearchDatastore_Task, 'duration_secs': 0.011621} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.099629] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.100336] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.100336] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.100665] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 737.100878] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 737.102830] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 737.102830] env[62000]: value = "task-882009" [ 737.102830] env[62000]: _type = "Task" [ 737.102830] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.102830] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08b5ea6b-3440-4a36-aa0d-895c5b3a5d51 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.111227] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 737.111227] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5212d280-ea5e-e375-0bc7-897d0fb54127" [ 737.111227] env[62000]: _type = "Task" [ 737.111227] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.114080] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882009, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.121379] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5212d280-ea5e-e375-0bc7-897d0fb54127, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.302304] env[62000]: DEBUG oslo_vmware.api [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882004, 'name': RemoveSnapshot_Task, 'duration_secs': 0.842967} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.302607] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 737.302866] env[62000]: INFO nova.compute.manager [None req-465634d5-17a1-4255-bfc3-1f3c2bbe5163 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 18.08 seconds to snapshot the instance on the hypervisor. [ 737.455767] env[62000]: DEBUG oslo_vmware.api [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882005, 'name': PowerOnVM_Task, 'duration_secs': 0.63141} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.456375] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 737.456694] env[62000]: INFO nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Took 9.49 seconds to spawn the instance on the hypervisor. [ 737.456925] env[62000]: DEBUG nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 737.457925] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b14c65-4017-4503-b759-5381ae9dce13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.581285] env[62000]: DEBUG nova.scheduler.client.report [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.616330] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882009, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484368} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.619856] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4e617bb5-84e4-4505-99e3-61289826f511/4e617bb5-84e4-4505-99e3-61289826f511.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 737.620248] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 737.621373] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48095b67-75fd-4c1c-9f13-48f64b5e97a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.629920] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5212d280-ea5e-e375-0bc7-897d0fb54127, 'name': SearchDatastore_Task, 'duration_secs': 0.017872} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.632598] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 737.632598] env[62000]: value = "task-882010" [ 737.632598] env[62000]: _type = "Task" [ 737.632598] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.632878] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f3056f6-c93a-426b-bcce-b74a63c57b0b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.642443] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 737.642443] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5250bd39-b07d-0d28-4501-f5c0973c8a5d" [ 737.642443] env[62000]: _type = "Task" [ 737.642443] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.645986] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882010, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.653235] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5250bd39-b07d-0d28-4501-f5c0973c8a5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.674118] env[62000]: DEBUG nova.compute.manager [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 737.674328] env[62000]: DEBUG nova.compute.manager [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing instance network info cache due to event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 737.675142] env[62000]: DEBUG oslo_concurrency.lockutils [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.675337] env[62000]: DEBUG oslo_concurrency.lockutils [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.675515] env[62000]: DEBUG nova.network.neutron [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 737.805637] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Updated VIF entry in instance network info cache for port e01b6797-8122-495f-b322-2212f250481b. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 737.806009] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Updating instance_info_cache with network_info: [{"id": "e01b6797-8122-495f-b322-2212f250481b", "address": "fa:16:3e:fa:bb:6e", "network": {"id": "a4de1643-552e-424b-ba9b-17ff06d7270d", "bridge": "br-int", "label": "tempest-AttachInterfacesV270Test-1878559773-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "087cacaac4d64fcc85cc3fcad3ee9c2e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2b6c45fd-e930-495a-9cb7-df84eda443b1", "external-id": "nsx-vlan-transportzone-407", "segmentation_id": 407, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape01b6797-81", "ovs_interfaceid": "e01b6797-8122-495f-b322-2212f250481b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.987667] env[62000]: INFO nova.compute.manager [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Took 25.97 seconds to build instance. [ 738.092115] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.509s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.093276] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 738.096019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.128s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.097647] env[62000]: INFO nova.compute.claims [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.149249] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882010, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.09561} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.153942] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 738.155813] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08b759fb-a73f-4f02-8de1-b728c37810a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.167218] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5250bd39-b07d-0d28-4501-f5c0973c8a5d, 'name': SearchDatastore_Task, 'duration_secs': 0.038887} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.181853] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.182287] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 738.199427] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Reconfiguring VM instance instance-0000002d to attach disk [datastore2] 4e617bb5-84e4-4505-99e3-61289826f511/4e617bb5-84e4-4505-99e3-61289826f511.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 738.199844] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.200158] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.200495] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-be5a18eb-4c11-49fb-af48-b8325b3064f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.204141] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c695a117-7fe1-4db5-9b74-8e0453c00c5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.229817] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-06b89289-d930-4117-b426-0d26790bfa02 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.240504] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 738.240504] env[62000]: value = "task-882012" [ 738.240504] env[62000]: _type = "Task" [ 738.240504] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.242312] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 738.242312] env[62000]: value = "task-882011" [ 738.242312] env[62000]: _type = "Task" [ 738.242312] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.250150] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.250444] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 738.255329] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6279acf5-5ecd-4bbf-b2a7-e3209b1cca66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.266307] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882011, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.272626] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 738.272626] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520259b4-25a9-b7e7-3a83-8042ee3146ee" [ 738.272626] env[62000]: _type = "Task" [ 738.272626] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.273462] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.288415] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520259b4-25a9-b7e7-3a83-8042ee3146ee, 'name': SearchDatastore_Task, 'duration_secs': 0.013171} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.289683] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-30287325-1190-4a83-b251-e3363dfefca5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.296781] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 738.296781] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520cd56d-6429-f721-72f1-a28ca2d4c6c4" [ 738.296781] env[62000]: _type = "Task" [ 738.296781] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.310110] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Releasing lock "refresh_cache-af49714d-8e50-4159-96a5-cf8f70580471" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.310526] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Received event network-changed-892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 738.310822] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Refreshing instance network info cache due to event network-changed-892944d2-5b6b-4a93-b975-f9e1f8104b1e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 738.311200] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquiring lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.311455] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquired lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.311746] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Refreshing network info cache for port 892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 738.313464] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520cd56d-6429-f721-72f1-a28ca2d4c6c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009188} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.314273] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.314565] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] af49714d-8e50-4159-96a5-cf8f70580471/af49714d-8e50-4159-96a5-cf8f70580471.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 738.314850] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1e73752f-764d-47c2-a059-277dc6831796 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.322727] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 738.322727] env[62000]: value = "task-882013" [ 738.322727] env[62000]: _type = "Task" [ 738.322727] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.332779] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882013, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.490249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-57f7fdd6-ea8a-4954-8a24-aea03786e6c7 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.332s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.606195] env[62000]: DEBUG nova.compute.utils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 738.608139] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 738.608369] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 738.616044] env[62000]: DEBUG nova.network.neutron [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updated VIF entry in instance network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 738.616550] env[62000]: DEBUG nova.network.neutron [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.679759] env[62000]: DEBUG nova.policy [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79730e91610c4c598019b122b4f41418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6106a70abdad4111977eebbd82434337', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 738.760961] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882012, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.765555] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882011, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502069} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 738.769575] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 738.769575] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 738.769575] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5d86a96f-c131-4e82-8ccc-a362ae4de293 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.783266] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 738.783266] env[62000]: value = "task-882014" [ 738.783266] env[62000]: _type = "Task" [ 738.783266] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 738.795148] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882014, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.843848] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882013, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 738.983078] env[62000]: INFO nova.compute.manager [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Rebuilding instance [ 738.991181] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.991727] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 738.995089] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 739.059182] env[62000]: DEBUG nova.compute.manager [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 739.059182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eca0ae-02b0-4407-8aca-9cfb3fe5ac2d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.115159] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 739.126027] env[62000]: DEBUG oslo_concurrency.lockutils [req-ea4ed0f2-942c-4c3d-8583-177e272b44e8 req-551f4104-5295-4d72-a422-b7d513146129 service nova] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.256214] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882012, 'name': ReconfigVM_Task, 'duration_secs': 0.622454} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.259050] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Reconfigured VM instance instance-0000002d to attach disk [datastore2] 4e617bb5-84e4-4505-99e3-61289826f511/4e617bb5-84e4-4505-99e3-61289826f511.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.259809] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d9ec456-44dc-4eca-bb75-e6075a8a2e58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.268039] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 739.268039] env[62000]: value = "task-882015" [ 739.268039] env[62000]: _type = "Task" [ 739.268039] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.279695] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882015, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.293433] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882014, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084951} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.293795] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 739.294613] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-085b2634-3115-4506-afe3-babbf748468f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.324809] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Reconfiguring VM instance instance-0000002f to attach disk [datastore2] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 739.329410] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cad937e4-54cf-4c3a-90c2-84955fd6df3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.346986] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Updated VIF entry in instance network info cache for port 892944d2-5b6b-4a93-b975-f9e1f8104b1e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 739.346986] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Updating instance_info_cache with network_info: [{"id": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "address": "fa:16:3e:47:3b:52", "network": {"id": "941616e2-425d-4828-847e-c1f3f4b49d34", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-381389191-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d2a32e7597d640c0a17818e965dbf4d3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "ba4f6497-e2b4-43b5-9819-6927865ae974", "external-id": "nsx-vlan-transportzone-112", "segmentation_id": 112, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap892944d2-5b", "ovs_interfaceid": "892944d2-5b6b-4a93-b975-f9e1f8104b1e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.357382] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882013, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.819864} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.358428] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] af49714d-8e50-4159-96a5-cf8f70580471/af49714d-8e50-4159-96a5-cf8f70580471.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 739.358591] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 739.359295] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 739.359295] env[62000]: value = "task-882016" [ 739.359295] env[62000]: _type = "Task" [ 739.359295] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.359295] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7a17a2e0-1303-47ec-b25a-c4693878bdca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.370718] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882016, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.375766] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 739.375766] env[62000]: value = "task-882017" [ 739.375766] env[62000]: _type = "Task" [ 739.375766] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.385596] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Successfully created port: adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 739.393358] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882017, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.407856] env[62000]: DEBUG nova.compute.manager [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 739.408111] env[62000]: DEBUG nova.compute.manager [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing instance network info cache due to event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 739.408361] env[62000]: DEBUG oslo_concurrency.lockutils [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.408531] env[62000]: DEBUG oslo_concurrency.lockutils [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.408720] env[62000]: DEBUG nova.network.neutron [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 739.509570] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 739.509842] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 739.510179] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Rebuilding the list of instances to heal {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 739.530649] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-850c2aa8-2a5a-471c-9442-6c83579ad1aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.534035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.539100] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1fb48f-7504-4e5d-a14c-bfc0b23d5d61 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.571785] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0b5b2f-6a50-429e-904a-f717cc606d64 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.575811] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 739.576110] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4827144d-d65d-4d7e-a0e5-c21c3d2a7614 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.583867] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1195d1ac-6b1f-4759-be28-291ec7cc84ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.588106] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 739.588106] env[62000]: value = "task-882018" [ 739.588106] env[62000]: _type = "Task" [ 739.588106] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.600852] env[62000]: DEBUG nova.compute.provider_tree [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.607603] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882018, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.788818] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882015, 'name': Rename_Task, 'duration_secs': 0.137866} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.790845] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 739.790845] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-93b87bfd-5c73-4739-95dd-3e6abb3f9a26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.797209] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 739.797209] env[62000]: value = "task-882019" [ 739.797209] env[62000]: _type = "Task" [ 739.797209] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.805730] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.854598] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Releasing lock "refresh_cache-210af329-4cdb-4c3f-9e82-e72a2ea79421" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.854598] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 739.854598] env[62000]: DEBUG nova.compute.manager [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing instance network info cache due to event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 739.854598] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.871453] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882016, 'name': ReconfigVM_Task, 'duration_secs': 0.265083} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.871709] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Reconfigured VM instance instance-0000002f to attach disk [datastore2] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 739.872408] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5cf52806-5562-465c-aaff-fd1be35ee45b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.878604] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 739.878604] env[62000]: value = "task-882020" [ 739.878604] env[62000]: _type = "Task" [ 739.878604] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.888947] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.890077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.890077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.890077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.890400] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.892702] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882017, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.129768} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 739.896911] env[62000]: INFO nova.compute.manager [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Terminating instance [ 739.899327] env[62000]: DEBUG nova.compute.manager [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.899529] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 739.900011] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 739.900300] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882020, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.901205] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-affe8cf2-307e-4036-b8c0-8af05c604a18 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.904709] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa10707-2c53-425b-b9d6-7c418a2dfcef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.915138] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 739.924216] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8908f83-a81b-4454-83d3-69b9697e7e38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.935733] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Reconfiguring VM instance instance-0000002e to attach disk [datastore2] af49714d-8e50-4159-96a5-cf8f70580471/af49714d-8e50-4159-96a5-cf8f70580471.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 739.935733] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9c1fcaf6-fcd4-43c5-8760-e249f754d984 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.956089] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 739.956089] env[62000]: value = "task-882021" [ 739.956089] env[62000]: _type = "Task" [ 739.956089] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.956944] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 739.956944] env[62000]: value = "task-882022" [ 739.956944] env[62000]: _type = "Task" [ 739.956944] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.978138] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882022, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.978472] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.019861] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Skipping network cache update for instance because it is being deleted. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9943}} [ 740.020231] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 740.020619] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 740.020619] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 740.020955] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 740.020955] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 740.057239] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 740.057427] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquired lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.057553] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Forcefully refreshing network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 740.057708] env[62000]: DEBUG nova.objects.instance [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lazy-loading 'info_cache' on Instance uuid 5e80a5c8-030c-4ad8-90c4-26136fa39d71 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 740.100057] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882018, 'name': PowerOffVM_Task, 'duration_secs': 0.456426} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.100468] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 740.100703] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.105045] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d623c323-f483-4187-a1a6-8e33cd8dae27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.105228] env[62000]: DEBUG nova.scheduler.client.report [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.114015] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 740.114015] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b59451f-db04-4eaf-8209-69e939cdf429 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.130835] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 740.162554] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 740.162812] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 740.162960] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 740.163457] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 740.163457] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 740.163457] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 740.163764] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 740.163799] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 740.166021] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 740.166021] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 740.166021] env[62000]: DEBUG nova.virt.hardware [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 740.166021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e2315b6-e61f-42b6-85cf-e24a42c4ebb5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.174253] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8282edac-f400-42d7-9c87-108d2f31d45a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.190156] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 740.190395] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 740.190582] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore2] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.190843] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-66e2de47-f2af-4c02-8d60-eb9166f520a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.196603] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 740.196603] env[62000]: value = "task-882024" [ 740.196603] env[62000]: _type = "Task" [ 740.196603] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.204403] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882024, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.308253] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882019, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.350029] env[62000]: DEBUG nova.network.neutron [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updated VIF entry in instance network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 740.350502] env[62000]: DEBUG nova.network.neutron [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.248", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.391289] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882020, 'name': Rename_Task, 'duration_secs': 0.190767} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.391596] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 740.391854] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e97fad02-d7d0-4012-a899-458eead7108f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.397455] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 740.397455] env[62000]: value = "task-882025" [ 740.397455] env[62000]: _type = "Task" [ 740.397455] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.406902] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882025, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.470604] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882022, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.475457] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882021, 'name': PowerOffVM_Task, 'duration_secs': 0.188715} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.475457] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 740.475457] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 740.475457] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9858964f-9340-448c-8f01-2e20a2bf766a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.518021] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.518384] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.549165] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 740.552022] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 740.552022] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Deleting the datastore file [datastore1] 1298037f-9cb7-4e2b-b70f-2ab24efe9b91 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.552022] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9e4df823-2045-4427-9634-1f6166b4f0da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.559956] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for the task: (returnval){ [ 740.559956] env[62000]: value = "task-882027" [ 740.559956] env[62000]: _type = "Task" [ 740.559956] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.572297] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882027, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.610969] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.515s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.611559] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.615207] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.327s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.616738] env[62000]: INFO nova.compute.claims [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.708959] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882024, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367573} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.709447] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.709840] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 740.710579] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 740.811277] env[62000]: DEBUG oslo_vmware.api [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882019, 'name': PowerOnVM_Task, 'duration_secs': 0.547705} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.811941] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 740.812184] env[62000]: INFO nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Took 10.29 seconds to spawn the instance on the hypervisor. [ 740.812447] env[62000]: DEBUG nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 740.813360] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cebd78a-7a84-4a70-9926-8d21e4dacbb5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.853059] env[62000]: DEBUG oslo_concurrency.lockutils [req-0dbdb1a4-0087-4aa0-b0b3-e4dc08818412 req-3c3bb463-4047-4f77-a950-5ded3295d70d service nova] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 740.853570] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 740.853761] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 740.910454] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882025, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.970497] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882022, 'name': ReconfigVM_Task, 'duration_secs': 0.756427} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.970743] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Reconfigured VM instance instance-0000002e to attach disk [datastore2] af49714d-8e50-4159-96a5-cf8f70580471/af49714d-8e50-4159-96a5-cf8f70580471.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 740.971406] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fffd9668-035d-4192-9db1-2c149433a102 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.978860] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 740.978860] env[62000]: value = "task-882028" [ 740.978860] env[62000]: _type = "Task" [ 740.978860] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.993151] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882028, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.074931] env[62000]: DEBUG oslo_vmware.api [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Task: {'id': task-882027, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259884} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.074931] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.074931] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 741.074931] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 741.074931] env[62000]: INFO nova.compute.manager [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 1.18 seconds to destroy the instance on the hypervisor. [ 741.075149] env[62000]: DEBUG oslo.service.loopingcall [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.075342] env[62000]: DEBUG nova.compute.manager [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 741.075464] env[62000]: DEBUG nova.network.neutron [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 741.121648] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.123564] env[62000]: DEBUG nova.compute.utils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.125576] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.127018] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 741.254028] env[62000]: DEBUG nova.policy [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 741.337919] env[62000]: INFO nova.compute.manager [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Took 25.63 seconds to build instance. [ 741.411038] env[62000]: DEBUG oslo_vmware.api [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882025, 'name': PowerOnVM_Task, 'duration_secs': 0.629803} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.411559] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 741.411816] env[62000]: INFO nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Took 5.78 seconds to spawn the instance on the hypervisor. [ 741.412081] env[62000]: DEBUG nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 741.413240] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a024c65-bdce-439d-bf8e-028f8376eaec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.473206] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Successfully updated port: adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 741.488899] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882028, 'name': Rename_Task, 'duration_secs': 0.18455} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.489200] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 741.489560] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a4b401a1-ad46-49db-a88f-9ee22cd584be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.496166] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 741.496166] env[62000]: value = "task-882029" [ 741.496166] env[62000]: _type = "Task" [ 741.496166] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.506731] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.630602] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.636378] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updated VIF entry in instance network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 741.636378] env[62000]: DEBUG nova.network.neutron [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.756342] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 741.757375] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 741.757375] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 741.757375] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 741.757375] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 741.757559] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 741.757657] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 741.757819] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 741.757988] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 741.758165] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 741.758346] env[62000]: DEBUG nova.virt.hardware [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 741.759783] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486f7787-9f61-4eb8-885d-e128744d3c50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.775710] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fc390a-e92d-4642-accb-9193c4beba75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.795548] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e4:58:db', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 741.806151] env[62000]: DEBUG oslo.service.loopingcall [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.806151] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 741.806517] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e1ed5709-97b2-4ced-a011-aed6f56aff25 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.821537] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 741.831883] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 741.831883] env[62000]: value = "task-882030" [ 741.831883] env[62000]: _type = "Task" [ 741.831883] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.839643] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0824eef1-d1d4-4b19-9e98-ba0be23b94a0 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.024s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.847255] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882030, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.931375] env[62000]: INFO nova.compute.manager [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Took 21.39 seconds to build instance. [ 741.951022] env[62000]: DEBUG nova.compute.manager [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 741.951252] env[62000]: DEBUG nova.compute.manager [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing instance network info cache due to event network-changed-1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 741.951442] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Acquiring lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.979778] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.979990] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.980154] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 742.012221] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882029, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.061207] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14f4ae4-8055-4c66-a4b1-38313d034935 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.071171] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a645d71-7b82-4945-bb0d-d08dd9a645fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.106041] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-486f65cb-1220-4d5f-94c1-4492eeb3ea71 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.114278] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fce29423-353d-4f6d-8a0d-890481fa94fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.128378] env[62000]: DEBUG nova.compute.provider_tree [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.141825] env[62000]: DEBUG oslo_concurrency.lockutils [req-3407b051-26e6-472c-8230-7af4fb707bf1 req-14cd5b0b-2160-40c8-aa59-2b5583501ab3 service nova] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.142638] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Acquired lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.142884] env[62000]: DEBUG nova.network.neutron [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Refreshing network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 742.220146] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Successfully created port: 7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.249833] env[62000]: DEBUG nova.network.neutron [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.327021] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Releasing lock "refresh_cache-5e80a5c8-030c-4ad8-90c4-26136fa39d71" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.327021] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Updated the network info_cache for instance {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 742.327247] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.327405] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.327548] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.327709] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.327837] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.328037] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.328174] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 742.328317] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 742.341968] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882030, 'name': CreateVM_Task, 'duration_secs': 0.444562} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.342763] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 742.343578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.343703] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.343909] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 742.344458] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4daf1dab-37a9-4939-8d07-a2335f864fc4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.347164] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 742.350872] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 742.350872] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c31183-d981-5e39-cd48-21471372a439" [ 742.350872] env[62000]: _type = "Task" [ 742.350872] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.359343] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c31183-d981-5e39-cd48-21471372a439, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.433982] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c57420fc-a959-4f2e-baea-d37267350099 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.998s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.514546] env[62000]: DEBUG oslo_vmware.api [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882029, 'name': PowerOnVM_Task, 'duration_secs': 0.643401} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.514546] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 742.514546] env[62000]: INFO nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Took 9.43 seconds to spawn the instance on the hypervisor. [ 742.514546] env[62000]: DEBUG nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 742.515758] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ebab18-66aa-4473-9284-91c57c22b0a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.528182] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.631648] env[62000]: DEBUG nova.scheduler.client.report [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.643782] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.667512] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.667900] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.668163] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.668365] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.668543] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.671397] env[62000]: INFO nova.compute.manager [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Terminating instance [ 742.675166] env[62000]: DEBUG nova.compute.manager [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 742.675390] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 742.677014] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d48e2d3d-934a-4fdc-8167-981fcd15697b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.686174] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.686401] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.686563] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.686750] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.686901] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.689036] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.689281] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.689428] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.689610] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.689913] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.690011] env[62000]: DEBUG nova.virt.hardware [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.691024] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a58e774-4781-40a1-bbb6-d69d8ec5ffb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.699204] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 742.699204] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f8f8058-df88-4485-bf05-ce6d4244b529 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.703135] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0ddfee-2f3b-475e-96f4-62aa6b5658f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.715014] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 742.715014] env[62000]: value = "task-882031" [ 742.715014] env[62000]: _type = "Task" [ 742.715014] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.735865] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-882031, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.752715] env[62000]: INFO nova.compute.manager [-] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Took 1.68 seconds to deallocate network for instance. [ 742.792459] env[62000]: DEBUG nova.network.neutron [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.831581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.870022] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c31183-d981-5e39-cd48-21471372a439, 'name': SearchDatastore_Task, 'duration_secs': 0.029181} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.870022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.870022] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 742.870022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.870022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.870022] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 742.870022] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f0b23c21-7221-4ad9-8448-d96700b6b1c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.877123] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.883021] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 742.883021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 742.883021] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8532f5d-3587-4212-9c95-f2a6d6581aae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.887526] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 742.887526] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52efedce-c6c1-0d10-2161-679ed75171b7" [ 742.887526] env[62000]: _type = "Task" [ 742.887526] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.897569] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52efedce-c6c1-0d10-2161-679ed75171b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.936655] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.010022] env[62000]: DEBUG nova.network.neutron [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updated VIF entry in instance network info cache for port 1d7778e6-8a84-447d-a241-95ea69ab68ba. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 743.010690] env[62000]: DEBUG nova.network.neutron [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [{"id": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "address": "fa:16:3e:3d:56:f3", "network": {"id": "d0e3b15e-da61-4b3e-85bb-f6f89583f383", "bridge": "br-int", "label": "tempest-ServerRescueTestJSONUnderV235-563613680-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "c051bb06d4c04417902263736e94a8b0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "365ac5b1-6d83-4dfe-887f-60574d7f6124", "external-id": "nsx-vlan-transportzone-138", "segmentation_id": 138, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1d7778e6-8a", "ovs_interfaceid": "1d7778e6-8a84-447d-a241-95ea69ab68ba", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.035010] env[62000]: INFO nova.compute.manager [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Took 26.17 seconds to build instance. [ 743.142763] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.528s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.143231] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 743.145835] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.573s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.146077] env[62000]: DEBUG nova.objects.instance [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lazy-loading 'resources' on Instance uuid 415cc4a5-7610-4678-971d-cd00a0e8b54d {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 743.223323] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-882031, 'name': PowerOffVM_Task, 'duration_secs': 0.222349} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.223899] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 743.224118] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 743.224378] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-932282f3-be0a-418f-b170-0ee7e9b55540 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.262352] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.296316] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.296791] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Instance network_info: |[{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 743.298260] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:98:64:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'adbea477-8af6-4cb5-82e6-1292553013f2', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 743.305966] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating folder: Project (6106a70abdad4111977eebbd82434337). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.306287] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 743.306488] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 743.306669] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Deleting the datastore file [datastore1] 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 743.306916] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2cf5be2-27cd-42b2-85d3-6972fe34162f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.309160] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aa301540-0e6f-4dba-acfd-0379108b83c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.316904] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for the task: (returnval){ [ 743.316904] env[62000]: value = "task-882034" [ 743.316904] env[62000]: _type = "Task" [ 743.316904] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.321334] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created folder: Project (6106a70abdad4111977eebbd82434337) in parent group-v201431. [ 743.321532] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating folder: Instances. Parent ref: group-v201496. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 743.322093] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d44b3ad7-f1be-485a-a3f1-fb3db8a3094e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.323854] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "4e617bb5-84e4-4505-99e3-61289826f511" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.324121] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.324348] env[62000]: INFO nova.compute.manager [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Rebooting instance [ 743.328847] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-882034, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.336386] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created folder: Instances in parent group-v201496. [ 743.336700] env[62000]: DEBUG oslo.service.loopingcall [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.337020] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 743.337020] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-80888582-1a2a-45b3-9651-0465a57e6c7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.359091] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 743.359091] env[62000]: value = "task-882036" [ 743.359091] env[62000]: _type = "Task" [ 743.359091] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.370613] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882036, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.398275] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52efedce-c6c1-0d10-2161-679ed75171b7, 'name': SearchDatastore_Task, 'duration_secs': 0.012669} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.399218] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7fae44e-c257-432a-8568-347adbb9b97a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.404938] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 743.404938] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d4c81c-f722-1447-a008-b7afcefcd820" [ 743.404938] env[62000]: _type = "Task" [ 743.404938] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.414215] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d4c81c-f722-1447-a008-b7afcefcd820, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.424551] env[62000]: INFO nova.compute.manager [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Rebuilding instance [ 743.464224] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.472986] env[62000]: DEBUG nova.compute.manager [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 743.473924] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa802942-4fd8-45d3-adff-8ed8800e9ea0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.513418] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Releasing lock "refresh_cache-64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.513703] env[62000]: DEBUG nova.compute.manager [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received event network-vif-plugged-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 743.514012] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Acquiring lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.514379] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.514609] env[62000]: DEBUG oslo_concurrency.lockutils [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.514873] env[62000]: DEBUG nova.compute.manager [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] No waiting events found dispatching network-vif-plugged-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 743.515065] env[62000]: WARNING nova.compute.manager [req-f8e40a2d-a708-474f-a8f2-b4a99ee6db4f req-76005bc2-aef0-4dde-9347-2f757bd4d899 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received unexpected event network-vif-plugged-adbea477-8af6-4cb5-82e6-1292553013f2 for instance with vm_state building and task_state spawning. [ 743.537459] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8dd7ae17-00fa-4dc1-b679-b6ae4e9a2fa5 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.068s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.649786] env[62000]: DEBUG nova.compute.utils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.654970] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.655111] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 743.683983] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "interface-af49714d-8e50-4159-96a5-cf8f70580471-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.684292] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "interface-af49714d-8e50-4159-96a5-cf8f70580471-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.684610] env[62000]: DEBUG nova.objects.instance [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lazy-loading 'flavor' on Instance uuid af49714d-8e50-4159-96a5-cf8f70580471 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 743.764174] env[62000]: DEBUG nova.policy [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '147e732bf4e94249abef07fd2090ea80', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '45a0491d5e13450bb11b99c4fba2807f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 743.829732] env[62000]: DEBUG oslo_vmware.api [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Task: {'id': task-882034, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.323321} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.831569] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 743.831830] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 743.831967] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 743.832162] env[62000]: INFO nova.compute.manager [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Took 1.16 seconds to destroy the instance on the hypervisor. [ 743.832404] env[62000]: DEBUG oslo.service.loopingcall [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 743.832776] env[62000]: DEBUG nova.compute.manager [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 743.832871] env[62000]: DEBUG nova.network.neutron [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 743.853943] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.854175] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquired lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.854420] env[62000]: DEBUG nova.network.neutron [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.871573] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882036, 'name': CreateVM_Task, 'duration_secs': 0.384254} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.871759] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 743.872517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.872696] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.873077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 743.873449] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d8d944ed-3bcb-4039-b14a-d5937aa2c708 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.883151] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 743.883151] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205a592-6bcc-5008-659b-559e18256955" [ 743.883151] env[62000]: _type = "Task" [ 743.883151] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.892817] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205a592-6bcc-5008-659b-559e18256955, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.918431] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d4c81c-f722-1447-a008-b7afcefcd820, 'name': SearchDatastore_Task, 'duration_secs': 0.012209} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 743.921182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.921478] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 743.921962] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bb28ba82-ad7f-47af-ab35-1dfcbbcf8cdf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.930280] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 743.930280] env[62000]: value = "task-882037" [ 743.930280] env[62000]: _type = "Task" [ 743.930280] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 743.943825] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882037, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.985385] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 743.987018] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-515f5367-4119-4c25-98f1-e3ad27034aea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.995094] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 743.995094] env[62000]: value = "task-882038" [ 743.995094] env[62000]: _type = "Task" [ 743.995094] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.008305] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882038, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.018296] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62aef42f-cc8c-401d-bb81-7348cb6aaaff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.027969] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95beffe-ee08-4969-82e7-9335557557ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.063129] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7f1744-6e13-4d01-8562-2368d1e159e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.072021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef672f9b-10f5-41a9-a805-b5bc73945621 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.088314] env[62000]: DEBUG nova.compute.provider_tree [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 744.155179] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 744.191639] env[62000]: DEBUG nova.objects.instance [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lazy-loading 'pci_requests' on Instance uuid af49714d-8e50-4159-96a5-cf8f70580471 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 744.400384] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205a592-6bcc-5008-659b-559e18256955, 'name': SearchDatastore_Task, 'duration_secs': 0.030938} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.400714] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.400967] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 744.401243] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.401397] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.401591] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 744.402235] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d027cff-f63b-4df4-90a8-810bf4a01585 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.420844] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 744.421172] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 744.422252] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c2803816-3322-4f69-a221-e6338678967f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.431388] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 744.431388] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5299144e-84ef-eeb6-7561-b594e2df9c04" [ 744.431388] env[62000]: _type = "Task" [ 744.431388] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.446636] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882037, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.450288] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5299144e-84ef-eeb6-7561-b594e2df9c04, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.510123] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882038, 'name': PowerOffVM_Task, 'duration_secs': 0.148401} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.510455] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 744.510683] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 744.511503] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047ba08f-d3c6-448c-816b-8e918553853b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.519324] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 744.519543] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4fd3857a-ffa2-43f7-aa1b-63477ed98566 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.548780] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 744.549319] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 744.549960] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Deleting the datastore file [datastore2] a9783a2a-a7d3-4fbe-9590-d530ac24fa82 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 744.550319] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e1f26b9a-d55d-49f6-9ea5-4665d9dd984d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.560299] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 744.560299] env[62000]: value = "task-882040" [ 744.560299] env[62000]: _type = "Task" [ 744.560299] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.573683] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882040, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.612116] env[62000]: ERROR nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] [req-2198a6bd-7d82-4d39-92c3-a7e3ed6da66a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2198a6bd-7d82-4d39-92c3-a7e3ed6da66a"}]} [ 744.615679] env[62000]: DEBUG nova.network.neutron [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.635299] env[62000]: DEBUG nova.compute.manager [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 744.635299] env[62000]: DEBUG nova.compute.manager [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing instance network info cache due to event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 744.635299] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Acquiring lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.635299] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Acquired lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.635299] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 744.637488] env[62000]: DEBUG nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 744.665813] env[62000]: DEBUG nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 744.666130] env[62000]: DEBUG nova.compute.provider_tree [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 744.680144] env[62000]: DEBUG nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 744.694956] env[62000]: DEBUG nova.objects.base [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 744.695191] env[62000]: DEBUG nova.network.neutron [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 744.700067] env[62000]: DEBUG nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 744.711954] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Successfully created port: ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 744.714736] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Successfully updated port: 7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 744.898114] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06066a81-e19d-441f-abe1-a91976231ad1 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "interface-af49714d-8e50-4159-96a5-cf8f70580471-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.213s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.949682] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5299144e-84ef-eeb6-7561-b594e2df9c04, 'name': SearchDatastore_Task, 'duration_secs': 0.050971} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.950229] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882037, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634262} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.953735] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 744.954037] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.954574] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00cce2ac-3fdf-48fa-9e55-31deacbc1345 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.957379] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4e62f3e1-65a0-4b3c-8e0d-11bd5b237e97 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.963816] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 744.963816] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526cdea8-2b12-6c56-faad-5cf37debafc3" [ 744.963816] env[62000]: _type = "Task" [ 744.963816] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.972142] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 744.972142] env[62000]: value = "task-882041" [ 744.972142] env[62000]: _type = "Task" [ 744.972142] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.980318] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526cdea8-2b12-6c56-faad-5cf37debafc3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.987317] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882041, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.077288] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882040, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.464024} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.077557] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 745.077750] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 745.078131] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 745.105093] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f936c73-4349-4558-b0eb-9a9d688ce99e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.118572] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97dcf478-fe59-41ea-a7de-c8e74be68ef3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.125264] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Releasing lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.128014] env[62000]: DEBUG nova.compute.manager [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 745.130026] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c57770-a6c6-43cf-bd38-34ace19c14bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.930076] env[62000]: DEBUG nova.network.neutron [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.933054] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 745.935496] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.935626] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.935968] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.950299] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f98519-26eb-410b-9c8d-67c25712df26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.970065] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882041, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.132537} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.972606] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:54:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='622b3a6c-bfbe-431a-b5c4-116aa70b0e48',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-862518541',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 745.972925] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 745.973067] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 745.973254] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 745.973399] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 745.973545] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 745.973750] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 745.973907] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 745.974086] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 745.974250] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 745.974420] env[62000]: DEBUG nova.virt.hardware [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 745.978289] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.979350] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae15c274-38ba-4989-91f5-1c686d06b2b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.982670] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526cdea8-2b12-6c56-faad-5cf37debafc3, 'name': SearchDatastore_Task, 'duration_secs': 0.019763} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.983989] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfcc04a7-cd1b-498e-ae2f-24e68f78238b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.987531] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.987762] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c/f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 745.990937] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f1764e76-f218-4135-b0c1-f059a0aec786 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.996847] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae396d3-a4df-48b7-8685-130e23934619 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.013503] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-504c3916-248e-42c5-a755-5798e663d821 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.024986] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Reconfiguring VM instance instance-0000002c to attach disk [datastore1] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 746.029356] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb5e66c3-f78a-48e3-aa10-aba355cca99e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.044626] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 746.044626] env[62000]: value = "task-882042" [ 746.044626] env[62000]: _type = "Task" [ 746.044626] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.054617] env[62000]: DEBUG nova.compute.provider_tree [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.069927] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 746.069927] env[62000]: value = "task-882043" [ 746.069927] env[62000]: _type = "Task" [ 746.069927] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.076215] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882042, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.082096] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882043, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.421527] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updated VIF entry in instance network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 746.421984] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.443317] env[62000]: INFO nova.compute.manager [-] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Took 2.61 seconds to deallocate network for instance. [ 746.507783] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.549844] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 746.550158] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 746.550394] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 746.550516] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 746.550663] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 746.550812] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 746.551110] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 746.551239] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 746.551402] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 746.551645] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 746.551737] env[62000]: DEBUG nova.virt.hardware [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 746.553039] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eace95da-cf09-4a28-a1c4-f213b397533c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.565994] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882042, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.568440] env[62000]: DEBUG nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.572773] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddba2210-de55-4981-a51d-5ac5cf82f834 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.594520] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 746.600966] env[62000]: DEBUG oslo.service.loopingcall [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.605177] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 746.605461] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882043, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.605672] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44c0fd5e-259c-403b-a2a6-b1d1fad0d8d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.625488] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 746.625488] env[62000]: value = "task-882044" [ 746.625488] env[62000]: _type = "Task" [ 746.625488] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.637999] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882044, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.784404] env[62000]: DEBUG nova.network.neutron [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.836480] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Successfully updated port: ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 746.925062] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Releasing lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.925541] env[62000]: DEBUG nova.compute.manager [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Received event network-vif-deleted-ff7bb956-76bd-4472-99b0-ae8305b3f7b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 746.925863] env[62000]: DEBUG nova.compute.manager [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 746.926173] env[62000]: DEBUG nova.compute.manager [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing instance network info cache due to event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 746.926514] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Acquiring lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.926763] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Acquired lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.927059] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 746.965397] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.973271] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c23d61-4c20-4eeb-b060-18f10d5fa0b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.982100] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Doing hard reboot of VM {{(pid=62000) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1064}} [ 746.982100] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-9c279670-e901-4c9d-bf5a-f1dbe24a77fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.988587] env[62000]: DEBUG oslo_vmware.api [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 746.988587] env[62000]: value = "task-882045" [ 746.988587] env[62000]: _type = "Task" [ 746.988587] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.999990] env[62000]: DEBUG oslo_vmware.api [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882045, 'name': ResetVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.011394] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-plugged-7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.011631] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.011848] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.012142] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.012361] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] No waiting events found dispatching network-vif-plugged-7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 747.012550] env[62000]: WARNING nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received unexpected event network-vif-plugged-7961bfcb-4d94-4279-bc3a-b46ca382b0cf for instance with vm_state building and task_state spawning. [ 747.012728] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Received event network-vif-deleted-1d7778e6-8a84-447d-a241-95ea69ab68ba {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.012927] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-changed-7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.013330] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing instance network info cache due to event network-changed-7961bfcb-4d94-4279-bc3a-b46ca382b0cf. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 747.013541] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.070751] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882042, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.620211} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.072608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c/f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 747.072608] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 747.072608] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-71d2d88e-3ee6-4a3d-b5bf-6febf4cf358e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.078422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 3.933s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.081320] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.519s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.081500] env[62000]: DEBUG nova.objects.instance [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lazy-loading 'resources' on Instance uuid 5e80a5c8-030c-4ad8-90c4-26136fa39d71 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 747.088305] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882043, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.088794] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 747.088794] env[62000]: value = "task-882046" [ 747.088794] env[62000]: _type = "Task" [ 747.088794] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.100376] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.107931] env[62000]: INFO nova.scheduler.client.report [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Deleted allocations for instance 415cc4a5-7610-4678-971d-cd00a0e8b54d [ 747.136812] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882044, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.290019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.290019] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Instance network_info: |[{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 747.290019] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.290019] env[62000]: DEBUG nova.network.neutron [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing network info cache for port 7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 747.290785] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:28:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a15de394-0367-4921-a5c1-6ac8615e3283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7961bfcb-4d94-4279-bc3a-b46ca382b0cf', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 747.301068] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Creating folder: Project (747546b09fd04a41b9c2df860a699186). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.304308] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-caa3b874-1c88-4e6c-b954-63621e1a0fcb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.317703] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Created folder: Project (747546b09fd04a41b9c2df860a699186) in parent group-v201431. [ 747.318374] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Creating folder: Instances. Parent ref: group-v201500. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 747.319559] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eeb0c348-686c-4109-988f-c0f7d0e83da8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.335032] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Created folder: Instances in parent group-v201500. [ 747.335032] env[62000]: DEBUG oslo.service.loopingcall [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 747.335032] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 747.335032] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83849923-50d2-4c42-8315-d6ae5df91671 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.351098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.351270] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.351401] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.361951] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 747.361951] env[62000]: value = "task-882049" [ 747.361951] env[62000]: _type = "Task" [ 747.361951] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.372590] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882049, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.501748] env[62000]: DEBUG oslo_vmware.api [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882045, 'name': ResetVM_Task, 'duration_secs': 0.102365} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.505024] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Did hard reboot of VM {{(pid=62000) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1068}} [ 747.505024] env[62000]: DEBUG nova.compute.manager [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.505024] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190a6825-eb66-4494-b6fe-f1b396b0a73a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.572216] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "af49714d-8e50-4159-96a5-cf8f70580471" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.572478] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.572725] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "af49714d-8e50-4159-96a5-cf8f70580471-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.572927] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.573130] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.578290] env[62000]: INFO nova.compute.manager [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Terminating instance [ 747.583056] env[62000]: DEBUG nova.compute.manager [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 747.583346] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 747.584217] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb536d1f-559a-40ca-9c94-5ef4ba49298f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.596138] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882043, 'name': ReconfigVM_Task, 'duration_secs': 1.344815} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.605143] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Reconfigured VM instance instance-0000002c to attach disk [datastore1] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2/596af5ab-1791-42ce-93d2-3e4f0a47dfa2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 747.605143] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-985ffeb8-3233-4118-a80a-9d37aec570cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.612553] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 747.617365] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e0a4e82-df8e-4204-b1be-6a3a016fd608 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.622335] env[62000]: DEBUG oslo_concurrency.lockutils [None req-952b5545-ac4b-432f-9f89-ab3651188c62 tempest-ServerDiagnosticsNegativeTest-1368844272 tempest-ServerDiagnosticsNegativeTest-1368844272-project-member] Lock "415cc4a5-7610-4678-971d-cd00a0e8b54d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.129s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.626765] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 747.626765] env[62000]: value = "task-882050" [ 747.626765] env[62000]: _type = "Task" [ 747.626765] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.627388] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081468} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.636041] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 747.641309] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d1aa3b1-07e9-4f6d-ba72-d713330065ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.644717] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 747.644717] env[62000]: value = "task-882051" [ 747.644717] env[62000]: _type = "Task" [ 747.644717] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.672206] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882044, 'name': CreateVM_Task, 'duration_secs': 0.713989} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.679229] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c/f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 747.679590] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882050, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.680479] env[62000]: DEBUG nova.network.neutron [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updated VIF entry in instance network info cache for port 7961bfcb-4d94-4279-bc3a-b46ca382b0cf. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.680830] env[62000]: DEBUG nova.network.neutron [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.684600] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 747.685185] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4c860014-42af-4959-b14e-f324565c2e63 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.706723] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.707188] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Received event network-vif-plugged-ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 747.707513] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Acquiring lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.707851] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.708245] env[62000]: DEBUG oslo_concurrency.lockutils [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.708494] env[62000]: DEBUG nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] No waiting events found dispatching network-vif-plugged-ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 747.708804] env[62000]: WARNING nova.compute.manager [req-b2735050-1d90-4f7b-a32a-329c1081b486 req-9635680b-188a-44f8-908f-8aa9b42c7c2a service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Received unexpected event network-vif-plugged-ddd21181-32c0-4898-bd09-7689d6976198 for instance with vm_state building and task_state spawning. [ 747.709476] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.709743] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.710253] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 747.710715] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882051, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.711380] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45f6a4f0-8ce9-4df7-8b78-5c0391347594 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.718420] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 747.718420] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526f85fb-977b-77e9-dec3-f8331f426bc7" [ 747.718420] env[62000]: _type = "Task" [ 747.718420] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.724462] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 747.724462] env[62000]: value = "task-882052" [ 747.724462] env[62000]: _type = "Task" [ 747.724462] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 747.736884] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526f85fb-977b-77e9-dec3-f8331f426bc7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.744020] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882052, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.873507] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882049, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.898850] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updated VIF entry in instance network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 747.899306] env[62000]: DEBUG nova.network.neutron [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.906900] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.025153] env[62000]: DEBUG oslo_concurrency.lockutils [None req-63eba2f1-d804-4b4c-be12-b1e667bb50b2 tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 4.701s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.057188] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85466b7-fdb8-490e-8ab7-caa1ef20a84f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.066210] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78c2cc0-6c81-4e5a-b897-27a3d6887c30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.103854] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94493d81-c7f0-47fa-8429-9b19f0abe5ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.113189] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0517b5dd-a779-429d-89cf-4b6c6547b5b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.128457] env[62000]: DEBUG nova.compute.provider_tree [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.141764] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882050, 'name': Rename_Task, 'duration_secs': 0.199412} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.142710] env[62000]: DEBUG nova.network.neutron [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.144139] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 748.144556] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bb89359b-6b13-4942-aef2-5009b25d7286 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.153369] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 748.153369] env[62000]: value = "task-882053" [ 748.153369] env[62000]: _type = "Task" [ 748.153369] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.160224] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882051, 'name': PowerOffVM_Task, 'duration_secs': 0.221348} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.160958] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 748.161125] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 748.161658] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-29986594-7e4c-485e-a2c5-500bfee1b77e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.166967] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882053, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.231024] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526f85fb-977b-77e9-dec3-f8331f426bc7, 'name': SearchDatastore_Task, 'duration_secs': 0.012589} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.232752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.233037] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.233291] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.233438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.234021] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.234021] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-24a12243-99bb-49c6-8f10-c4a59dea65f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.242264] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882052, 'name': ReconfigVM_Task, 'duration_secs': 0.326171} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.242493] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Reconfigured VM instance instance-00000030 to attach disk [datastore1] f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c/f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 748.243146] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ae4e4a2f-de0f-46c2-a3c5-01b80c87c970 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.246091] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.246279] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.247362] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b80fe1a5-83fe-403d-b8bb-1a3b3c15c417 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.254107] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 748.254107] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c92f96-52a5-5dad-7b24-dd8fd1fcaaf7" [ 748.254107] env[62000]: _type = "Task" [ 748.254107] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.262020] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 748.262020] env[62000]: value = "task-882055" [ 748.262020] env[62000]: _type = "Task" [ 748.262020] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.263190] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 748.263423] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 748.263603] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Deleting the datastore file [datastore2] af49714d-8e50-4159-96a5-cf8f70580471 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 748.264222] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30736fae-34fc-4480-8c93-3b636a35e6ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.271207] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c92f96-52a5-5dad-7b24-dd8fd1fcaaf7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.274390] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882055, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.275764] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for the task: (returnval){ [ 748.275764] env[62000]: value = "task-882056" [ 748.275764] env[62000]: _type = "Task" [ 748.275764] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.284245] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882056, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.372336] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882049, 'name': CreateVM_Task, 'duration_secs': 0.522522} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.372636] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 748.373363] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.373490] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.373789] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 748.374122] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-216d2882-a945-4d1b-aff5-b88d61381f38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.381630] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 748.381630] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cae7b4-7249-ee3d-3219-5c27baf22caf" [ 748.381630] env[62000]: _type = "Task" [ 748.381630] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.389839] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cae7b4-7249-ee3d-3219-5c27baf22caf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.403640] env[62000]: DEBUG oslo_concurrency.lockutils [req-4da6c683-582f-4d3b-9eb1-942a456c2783 req-13db7526-69f1-440d-9f99-38629b05b2d7 service nova] Releasing lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.631541] env[62000]: DEBUG nova.scheduler.client.report [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.645883] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.646751] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Instance network_info: |[{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 748.646751] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e8:24:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ddd21181-32c0-4898-bd09-7689d6976198', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 748.655212] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Creating folder: Project (45a0491d5e13450bb11b99c4fba2807f). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 748.656369] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0cb5a44b-31a3-46e8-882d-c6bb68d5b79b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.668164] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882053, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.674733] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Created folder: Project (45a0491d5e13450bb11b99c4fba2807f) in parent group-v201431. [ 748.674733] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Creating folder: Instances. Parent ref: group-v201503. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 748.674733] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3839de8e-6c90-46b6-ae24-c45932c10231 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.685724] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Created folder: Instances in parent group-v201503. [ 748.686019] env[62000]: DEBUG oslo.service.loopingcall [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.686258] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 748.686529] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f98eaee7-ac9b-4fa8-b260-1595d432d426 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.718381] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 748.718381] env[62000]: value = "task-882059" [ 748.718381] env[62000]: _type = "Task" [ 748.718381] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.727588] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882059, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.771236] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c92f96-52a5-5dad-7b24-dd8fd1fcaaf7, 'name': SearchDatastore_Task, 'duration_secs': 0.019037} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.772709] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8192cfc7-f771-4d2d-9e6a-e8ebe322d747 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.778638] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882055, 'name': Rename_Task, 'duration_secs': 0.283145} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.783064] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 748.784578] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-aa48ffaa-cb13-4fbc-8cfd-ae80153ac021 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.786469] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 748.786469] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52390fb6-7774-5114-fa58-5661f21a8dfa" [ 748.786469] env[62000]: _type = "Task" [ 748.786469] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.795552] env[62000]: DEBUG oslo_vmware.api [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Task: {'id': task-882056, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.267978} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.796740] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 748.797135] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 748.797408] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 748.798033] env[62000]: INFO nova.compute.manager [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Took 1.21 seconds to destroy the instance on the hypervisor. [ 748.798627] env[62000]: DEBUG oslo.service.loopingcall [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.798974] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 748.798974] env[62000]: value = "task-882060" [ 748.798974] env[62000]: _type = "Task" [ 748.798974] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.803162] env[62000]: DEBUG nova.compute.manager [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.803753] env[62000]: DEBUG nova.network.neutron [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 748.805430] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52390fb6-7774-5114-fa58-5661f21a8dfa, 'name': SearchDatastore_Task, 'duration_secs': 0.013081} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.809955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.810283] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 748.811026] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-388085c7-74a6-426a-8d19-c59132ec20e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.819976] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882060, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.822495] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 748.822495] env[62000]: value = "task-882061" [ 748.822495] env[62000]: _type = "Task" [ 748.822495] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.832074] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882061, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 748.897599] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cae7b4-7249-ee3d-3219-5c27baf22caf, 'name': SearchDatastore_Task, 'duration_secs': 0.014189} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 748.897970] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.898292] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 748.898593] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.900732] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.900732] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 748.900732] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a2f5491-6285-4c1a-99fa-0a4e09958b93 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.910766] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 748.911192] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 748.914253] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-893fec8e-78bf-4da2-8b6b-34c3c0bff22e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.921852] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 748.921852] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52eb1475-a737-3ec8-f905-c447a28f0f78" [ 748.921852] env[62000]: _type = "Task" [ 748.921852] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 748.931273] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52eb1475-a737-3ec8-f905-c447a28f0f78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.049753] env[62000]: DEBUG nova.compute.manager [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Received event network-changed-ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.050148] env[62000]: DEBUG nova.compute.manager [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Refreshing instance network info cache due to event network-changed-ddd21181-32c0-4898-bd09-7689d6976198. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.050242] env[62000]: DEBUG oslo_concurrency.lockutils [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.050352] env[62000]: DEBUG oslo_concurrency.lockutils [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.050521] env[62000]: DEBUG nova.network.neutron [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Refreshing network info cache for port ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.137138] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.055s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.138630] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.508s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.140589] env[62000]: INFO nova.compute.claims [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.167265] env[62000]: DEBUG oslo_vmware.api [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882053, 'name': PowerOnVM_Task, 'duration_secs': 0.714883} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.167601] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 749.167866] env[62000]: DEBUG nova.compute.manager [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.168755] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0c41b4-e67f-4567-a3e8-a5fb41bdbf07 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.172876] env[62000]: INFO nova.scheduler.client.report [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Deleted allocations for instance 5e80a5c8-030c-4ad8-90c4-26136fa39d71 [ 749.230032] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882059, 'name': CreateVM_Task, 'duration_secs': 0.400397} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.230032] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 749.230773] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.230916] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.231269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 749.231568] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-103adfb5-4419-47e5-a6ea-33645410b7e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.239471] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 749.239471] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcce1c-06ce-8c32-5797-e0873313e1bd" [ 749.239471] env[62000]: _type = "Task" [ 749.239471] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.252365] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcce1c-06ce-8c32-5797-e0873313e1bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.317906] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882060, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.336452] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882061, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.432667] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52eb1475-a737-3ec8-f905-c447a28f0f78, 'name': SearchDatastore_Task, 'duration_secs': 0.011694} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.434684] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ccd98ec3-fb36-4524-b9f0-5c0e164d809a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.440615] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 749.440615] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5275ed21-812a-5f21-d375-c8ac22e8160c" [ 749.440615] env[62000]: _type = "Task" [ 749.440615] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.449287] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5275ed21-812a-5f21-d375-c8ac22e8160c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.496852] env[62000]: DEBUG nova.compute.manager [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 749.497058] env[62000]: DEBUG nova.compute.manager [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing instance network info cache due to event network-changed-bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 749.497299] env[62000]: DEBUG oslo_concurrency.lockutils [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] Acquiring lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.497645] env[62000]: DEBUG oslo_concurrency.lockutils [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] Acquired lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.497645] env[62000]: DEBUG nova.network.neutron [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Refreshing network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 749.681264] env[62000]: DEBUG oslo_concurrency.lockutils [None req-060245da-f8d2-44c2-9fce-29e7689fb93f tempest-ServerDiagnosticsV248Test-1086762981 tempest-ServerDiagnosticsV248Test-1086762981-project-member] Lock "5e80a5c8-030c-4ad8-90c4-26136fa39d71" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.973s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.703565] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.704479] env[62000]: DEBUG nova.network.neutron [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.751451] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcce1c-06ce-8c32-5797-e0873313e1bd, 'name': SearchDatastore_Task, 'duration_secs': 0.053054} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.751786] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.752074] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 749.752317] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.817051] env[62000]: DEBUG oslo_vmware.api [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882060, 'name': PowerOnVM_Task, 'duration_secs': 0.627507} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.817504] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 749.817814] env[62000]: INFO nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Took 9.69 seconds to spawn the instance on the hypervisor. [ 749.818147] env[62000]: DEBUG nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.819076] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b80c1f-abea-472f-8f47-de06f50773e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.842507] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882061, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.573125} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.842924] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 749.843744] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 749.844083] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1dfa6194-9427-4147-b843-cb442cb4f3aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.853354] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 749.853354] env[62000]: value = "task-882062" [ 749.853354] env[62000]: _type = "Task" [ 749.853354] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.866399] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882062, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.931198] env[62000]: DEBUG nova.network.neutron [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updated VIF entry in instance network info cache for port ddd21181-32c0-4898-bd09-7689d6976198. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 749.931198] env[62000]: DEBUG nova.network.neutron [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.956025] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5275ed21-812a-5f21-d375-c8ac22e8160c, 'name': SearchDatastore_Task, 'duration_secs': 0.030755} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 749.956025] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.956025] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 71192360-6ee6-4876-bf37-da987a09cbb2/71192360-6ee6-4876-bf37-da987a09cbb2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 749.956025] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.956025] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 749.956025] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d2f44fe4-28fe-4baf-86db-6573440c094e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.956844] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d506ad5b-4b0d-4de0-97ce-ec9e82b2d2e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.967479] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 749.967479] env[62000]: value = "task-882063" [ 749.967479] env[62000]: _type = "Task" [ 749.967479] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.977394] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 749.977394] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 749.977394] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-570f3cf1-fa69-4bd0-862d-703cf593e511 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.984235] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882063, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.988230] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 749.988230] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52268f90-0370-b971-5508-70cb379a5b17" [ 749.988230] env[62000]: _type = "Task" [ 749.988230] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.997717] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52268f90-0370-b971-5508-70cb379a5b17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.197379] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "4e617bb5-84e4-4505-99e3-61289826f511" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.197718] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.197943] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "4e617bb5-84e4-4505-99e3-61289826f511-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.198241] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.198448] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.202043] env[62000]: INFO nova.compute.manager [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Terminating instance [ 750.206050] env[62000]: DEBUG nova.compute.manager [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 750.206050] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 750.207482] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3ab2bf-9f0b-4674-be35-d8b59ca6d768 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.216343] env[62000]: INFO nova.compute.manager [-] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Took 1.41 seconds to deallocate network for instance. [ 750.231175] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 750.231175] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33db819a-262c-4733-a0cf-a821f1c736a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.241284] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 750.241284] env[62000]: value = "task-882064" [ 750.241284] env[62000]: _type = "Task" [ 750.241284] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.253236] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882064, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.352421] env[62000]: INFO nova.compute.manager [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Took 28.46 seconds to build instance. [ 750.371211] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882062, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074088} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.375254] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 750.377865] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0aa8cdf8-c584-4a8d-9c7b-46c298ade94b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.414207] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Reconfiguring VM instance instance-0000002f to attach disk [datastore1] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 750.415966] env[62000]: DEBUG nova.network.neutron [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updated VIF entry in instance network info cache for port bcf8ac20-4efb-4ad5-a14a-9548cfb59944. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 750.416229] env[62000]: DEBUG nova.network.neutron [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [{"id": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "address": "fa:16:3e:71:46:58", "network": {"id": "cf99ede7-0ed9-4d79-8c61-ac82a6d038eb", "bridge": "br-int", "label": "tempest-SecurityGroupsTestJSON-1063502461-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ad12e169457d491097ab311e05d8e30a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e1049e8-c06b-4c93-a9e1-2cbb530f3f95", "external-id": "nsx-vlan-transportzone-966", "segmentation_id": 966, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcf8ac20-4e", "ovs_interfaceid": "bcf8ac20-4efb-4ad5-a14a-9548cfb59944", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.422784] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-121620de-1830-4932-967f-0d1b920815e1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.442904] env[62000]: DEBUG oslo_concurrency.lockutils [req-108044ba-d630-4a7e-8d72-68f28d847565 req-ef02b608-c079-4b6c-99c0-e990d20e529b service nova] Releasing lock "refresh_cache-4e617bb5-84e4-4505-99e3-61289826f511" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.445091] env[62000]: DEBUG oslo_concurrency.lockutils [req-972bd272-41c0-40ac-ac16-d320f3966d70 req-faa14492-f54b-48dc-8e54-dd843224241c service nova] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.456000] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 750.456000] env[62000]: value = "task-882065" [ 750.456000] env[62000]: _type = "Task" [ 750.456000] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.467291] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882065, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.483304] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882063, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.500170] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52268f90-0370-b971-5508-70cb379a5b17, 'name': SearchDatastore_Task, 'duration_secs': 0.01924} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.504045] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc75c124-5e14-42bb-bd7c-9eb75769b6fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.511013] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 750.511013] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b6ecd6-cba3-9121-8d69-3d68221092b1" [ 750.511013] env[62000]: _type = "Task" [ 750.511013] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.522881] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b6ecd6-cba3-9121-8d69-3d68221092b1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.619797] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f024aa-f81d-4798-b41c-7502c2aa3a94 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.633658] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b4dc2d-3566-458e-94cb-2bf662acc3ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.677735] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f8cb40-8efb-4ffe-a449-fad4e780240d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.686872] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468447c8-138c-43fa-9133-d87d1537efb7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.704031] env[62000]: DEBUG nova.compute.provider_tree [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.733167] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.754175] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882064, 'name': PowerOffVM_Task, 'duration_secs': 0.241716} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.754464] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 750.754663] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 750.754928] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c20b92c2-7bf7-409e-81bc-93506e4d875b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.852949] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.852949] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.853184] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.853369] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.853597] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.856466] env[62000]: INFO nova.compute.manager [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Terminating instance [ 750.859025] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 750.859232] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 750.859418] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Deleting the datastore file [datastore2] 4e617bb5-84e4-4505-99e3-61289826f511 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.860549] env[62000]: DEBUG nova.compute.manager [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 750.861154] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 750.861154] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ae6139a3-6a31-4f96-8ef2-1a3f6ae3019a tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.475s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.861443] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-812d3319-590c-4c60-b722-e2a3594f22ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.864077] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f182c638-1676-4e5c-8d6a-b435310c7db7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.872803] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 750.874140] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-44d91751-a9ef-40ab-b076-ec918c645732 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.875842] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for the task: (returnval){ [ 750.875842] env[62000]: value = "task-882067" [ 750.875842] env[62000]: _type = "Task" [ 750.875842] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.882167] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 750.882167] env[62000]: value = "task-882068" [ 750.882167] env[62000]: _type = "Task" [ 750.882167] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.886036] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882067, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.897209] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882068, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.971072] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882065, 'name': ReconfigVM_Task, 'duration_secs': 0.488707} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.971072] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Reconfigured VM instance instance-0000002f to attach disk [datastore1] a9783a2a-a7d3-4fbe-9590-d530ac24fa82/a9783a2a-a7d3-4fbe-9590-d530ac24fa82.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 750.971072] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ba8be102-5f66-4a5e-8d9d-50087ee8c07a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.981215] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882063, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719887} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.982717] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 71192360-6ee6-4876-bf37-da987a09cbb2/71192360-6ee6-4876-bf37-da987a09cbb2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 750.982997] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 750.983543] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 750.983543] env[62000]: value = "task-882069" [ 750.983543] env[62000]: _type = "Task" [ 750.983543] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.983724] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6673e284-8aa7-4b95-9e18-8635dd4abc4f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.998258] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882069, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.998502] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 750.998502] env[62000]: value = "task-882070" [ 750.998502] env[62000]: _type = "Task" [ 750.998502] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.009538] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882070, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.022141] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b6ecd6-cba3-9121-8d69-3d68221092b1, 'name': SearchDatastore_Task, 'duration_secs': 0.064161} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.022563] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.023898] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 751.023898] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ae5e8130-8b16-4d57-9a46-d4836be3b5ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.031718] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 751.031718] env[62000]: value = "task-882071" [ 751.031718] env[62000]: _type = "Task" [ 751.031718] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.045716] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882071, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.208022] env[62000]: DEBUG nova.scheduler.client.report [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.218542] env[62000]: DEBUG nova.compute.manager [req-e0d0b2ce-6d64-4573-a3a2-3bb377fe1762 req-c9c00965-2582-4701-92da-10a089b2f512 service nova] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Received event network-vif-deleted-e01b6797-8122-495f-b322-2212f250481b {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 751.243725] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.244132] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.394279] env[62000]: DEBUG oslo_vmware.api [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Task: {'id': task-882067, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174483} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.397856] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.398040] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 751.398294] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 751.398652] env[62000]: INFO nova.compute.manager [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Took 1.19 seconds to destroy the instance on the hypervisor. [ 751.398966] env[62000]: DEBUG oslo.service.loopingcall [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.399253] env[62000]: DEBUG nova.compute.manager [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 751.399387] env[62000]: DEBUG nova.network.neutron [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.413341] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882068, 'name': PowerOffVM_Task, 'duration_secs': 0.328172} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.413341] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 751.413341] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 751.413341] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-46f5bd2c-3ac3-42aa-a9f4-bc02b5039a67 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.497359] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882069, 'name': Rename_Task, 'duration_secs': 0.178494} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.497735] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 751.498597] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3bf9990b-f48c-4202-b8ee-8d36e880a23c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.516641] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882070, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.473937} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.518300] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 751.518747] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 751.518747] env[62000]: value = "task-882073" [ 751.518747] env[62000]: _type = "Task" [ 751.518747] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.519564] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-969948a7-33cf-4c16-8198-5589845bcb75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.533899] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "ae343199-1006-473a-a47f-7983835bd60a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.534202] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.555370] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882073, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.565388] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfiguring VM instance instance-00000031 to attach disk [datastore1] 71192360-6ee6-4876-bf37-da987a09cbb2/71192360-6ee6-4876-bf37-da987a09cbb2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 751.567161] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-626605bc-325b-4b16-b4e5-2f963a01a0e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.586428] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882071, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509333} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.587146] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 751.587426] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 751.587800] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-14d187b9-8c10-49da-8ddb-128e6b94b241 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.593593] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 751.593593] env[62000]: value = "task-882074" [ 751.593593] env[62000]: _type = "Task" [ 751.593593] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.598589] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 751.598589] env[62000]: value = "task-882075" [ 751.598589] env[62000]: _type = "Task" [ 751.598589] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.605408] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882074, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.610619] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882075, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.714113] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.575s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.714936] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 751.719190] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 23.034s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.719593] env[62000]: DEBUG nova.objects.instance [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lazy-loading 'resources' on Instance uuid 72688992-333f-459d-9d05-f7c728961a6d {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 751.746958] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.035811] env[62000]: DEBUG oslo_vmware.api [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882073, 'name': PowerOnVM_Task, 'duration_secs': 0.515554} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.036610] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 752.036955] env[62000]: DEBUG nova.compute.manager [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 752.038592] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14c5cf45-4ccc-406f-830e-156567176c75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.046023] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 752.110821] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882075, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.114325] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882074, 'name': ReconfigVM_Task, 'duration_secs': 0.384394} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.114742] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfigured VM instance instance-00000031 to attach disk [datastore1] 71192360-6ee6-4876-bf37-da987a09cbb2/71192360-6ee6-4876-bf37-da987a09cbb2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 752.115277] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-061af72e-6e76-4489-9c71-684c7fb9593f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.124932] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 752.124932] env[62000]: value = "task-882076" [ 752.124932] env[62000]: _type = "Task" [ 752.124932] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.136261] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882076, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.223549] env[62000]: DEBUG nova.compute.utils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.228697] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 752.278783] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.332848] env[62000]: DEBUG nova.network.neutron [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.578069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.584360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.612675] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882075, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.969487} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.615469] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 752.616579] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1329d82d-d32b-45e3-9b15-cd4a54bdc15f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.641305] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 752.647420] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b40c1196-de73-4943-acb9-609952920087 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.665701] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 752.666078] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 752.666341] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore1] 596af5ab-1791-42ce-93d2-3e4f0a47dfa2 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 752.667520] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7410f0ff-f611-41b6-aeb3-6b7e3dd34cbc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.673742] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882076, 'name': Rename_Task, 'duration_secs': 0.420462} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.675592] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 752.676044] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 752.676044] env[62000]: value = "task-882077" [ 752.676044] env[62000]: _type = "Task" [ 752.676044] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.677171] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-05ff5ad1-ff3d-411f-b5dc-e7d90d2cfda9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.679517] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-272f6b6b-dd81-4b00-9c75-e1869dab151c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.688683] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 752.688683] env[62000]: value = "task-882078" [ 752.688683] env[62000]: _type = "Task" [ 752.688683] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.698249] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882077, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.704420] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 752.704420] env[62000]: value = "task-882079" [ 752.704420] env[62000]: _type = "Task" [ 752.704420] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.704710] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882078, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.705931] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90763362-88aa-4cac-8606-f3bf216c350c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.718682] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882079, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.748934] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 752.753687] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09aa13dd-d21f-4867-98ae-f89e3d3f4261 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.763939] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617f064b-740b-4080-86c9-2e22d5a05869 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.780764] env[62000]: DEBUG nova.compute.provider_tree [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.836392] env[62000]: INFO nova.compute.manager [-] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Took 1.44 seconds to deallocate network for instance. [ 752.891735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.891735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.891735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.891735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.891735] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.895256] env[62000]: INFO nova.compute.manager [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Terminating instance [ 752.897719] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "refresh_cache-a9783a2a-a7d3-4fbe-9590-d530ac24fa82" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.898079] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquired lock "refresh_cache-a9783a2a-a7d3-4fbe-9590-d530ac24fa82" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.898694] env[62000]: DEBUG nova.network.neutron [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 753.189550] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882077, 'name': ReconfigVM_Task, 'duration_secs': 0.426157} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.190033] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfigured VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 753.194673] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-67aa3d98-bf39-4399-9cb6-a6f4675e3519 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.205447] env[62000]: DEBUG oslo_vmware.api [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.226967} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.209770] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 753.210021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 753.210279] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 753.210381] env[62000]: INFO nova.compute.manager [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Took 2.35 seconds to destroy the instance on the hypervisor. [ 753.210668] env[62000]: DEBUG oslo.service.loopingcall [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.210931] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 753.210931] env[62000]: value = "task-882080" [ 753.210931] env[62000]: _type = "Task" [ 753.210931] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.211316] env[62000]: DEBUG nova.compute.manager [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.211417] env[62000]: DEBUG nova.network.neutron [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.221681] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882079, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.226721] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882080, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.264602] env[62000]: DEBUG nova.compute.manager [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 753.264813] env[62000]: DEBUG nova.compute.manager [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing instance network info cache due to event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 753.266182] env[62000]: DEBUG oslo_concurrency.lockutils [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] Acquiring lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.266408] env[62000]: DEBUG oslo_concurrency.lockutils [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] Acquired lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.266913] env[62000]: DEBUG nova.network.neutron [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 753.284295] env[62000]: DEBUG nova.scheduler.client.report [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 753.345528] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 753.430520] env[62000]: DEBUG nova.network.neutron [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.588288] env[62000]: DEBUG nova.network.neutron [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.716411] env[62000]: DEBUG oslo_vmware.api [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882079, 'name': PowerOnVM_Task, 'duration_secs': 1.014357} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.721378] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 753.721569] env[62000]: INFO nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Took 11.08 seconds to spawn the instance on the hypervisor. [ 753.721776] env[62000]: DEBUG nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 753.725697] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad7c4c4b-8c69-4c91-abcb-9b2abaca91ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.736137] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882080, 'name': Rename_Task, 'duration_secs': 0.241705} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.736677] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 753.736930] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f03bcd7-3ac7-4751-b3c8-4314cbe03da0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.745408] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 753.745408] env[62000]: value = "task-882081" [ 753.745408] env[62000]: _type = "Task" [ 753.745408] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.755452] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.769970] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 753.789359] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.070s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.791744] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.919s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.796197] env[62000]: INFO nova.compute.claims [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 753.810576] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 753.810861] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 753.811269] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 753.811431] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 753.811651] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 753.811828] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 753.812091] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 753.814019] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 753.814019] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 753.814019] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 753.814019] env[62000]: DEBUG nova.virt.hardware [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 753.814019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5f89ef0-3006-41b6-838a-0e102cfa4efe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.821918] env[62000]: INFO nova.scheduler.client.report [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Deleted allocations for instance 72688992-333f-459d-9d05-f7c728961a6d [ 753.829317] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c46c2dad-78df-45a5-9af1-3cb2064429d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.847816] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 753.853124] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Creating folder: Project (6aeccd1a0e3f400a99590f808d386ddb). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.857097] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-079ac537-0b7c-4e76-b386-90cdabdd2638 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.867104] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Created folder: Project (6aeccd1a0e3f400a99590f808d386ddb) in parent group-v201431. [ 753.867365] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Creating folder: Instances. Parent ref: group-v201506. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 753.867572] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d0f423c-7915-4541-a1f0-4eab106cc265 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.881293] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Created folder: Instances in parent group-v201506. [ 753.881293] env[62000]: DEBUG oslo.service.loopingcall [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.884023] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 753.884023] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-68d63817-ad32-4299-bc6d-aee0db434c16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.906958] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 753.906958] env[62000]: value = "task-882084" [ 753.906958] env[62000]: _type = "Task" [ 753.906958] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.916542] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882084, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.091357] env[62000]: DEBUG nova.network.neutron [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updated VIF entry in instance network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 754.091773] env[62000]: DEBUG nova.network.neutron [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.093512] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Releasing lock "refresh_cache-a9783a2a-a7d3-4fbe-9590-d530ac24fa82" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.093699] env[62000]: DEBUG nova.compute.manager [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 754.094613] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 754.094934] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f1060e6-d274-4819-981d-9fb184481d50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.104683] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 754.104944] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f898526-efbd-4225-b778-f4d4eab05a63 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.109319] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.109547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.115509] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 754.115509] env[62000]: value = "task-882085" [ 754.115509] env[62000]: _type = "Task" [ 754.115509] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.126143] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882085, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.169014] env[62000]: DEBUG nova.network.neutron [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.251473] env[62000]: INFO nova.compute.manager [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Took 32.31 seconds to build instance. [ 754.258964] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882081, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.338196] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b29b9616-ac37-4935-8947-ed7361676960 tempest-ServerMetadataNegativeTestJSON-78756092 tempest-ServerMetadataNegativeTestJSON-78756092-project-member] Lock "72688992-333f-459d-9d05-f7c728961a6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.069s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.418779] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882084, 'name': CreateVM_Task, 'duration_secs': 0.479974} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.419086] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 754.419544] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.419722] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.421360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 754.421360] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ed9695c-2f69-48aa-8a78-9bf0ae431ddb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.426553] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 754.426553] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d638f5-1693-c228-f5fd-4bfe9d5a2c88" [ 754.426553] env[62000]: _type = "Task" [ 754.426553] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.435885] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d638f5-1693-c228-f5fd-4bfe9d5a2c88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.595587] env[62000]: DEBUG oslo_concurrency.lockutils [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] Releasing lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.595884] env[62000]: DEBUG nova.compute.manager [req-0888eb87-9baf-47b9-8aa6-bc7d2a02c7fb req-bf377523-3df1-4a6f-adde-3f7e5fe513e8 service nova] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Received event network-vif-deleted-bcf8ac20-4efb-4ad5-a14a-9548cfb59944 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 754.612707] env[62000]: DEBUG nova.compute.utils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 754.626773] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882085, 'name': PowerOffVM_Task, 'duration_secs': 0.198214} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.627072] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 754.627271] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 754.627530] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-801eb39b-986b-41dc-938a-db9a5ace7e98 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.657260] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 754.657515] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 754.657709] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Deleting the datastore file [datastore1] a9783a2a-a7d3-4fbe-9590-d530ac24fa82 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 754.657974] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-050f9a9d-5452-4faa-9887-a02d80e978d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.665520] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for the task: (returnval){ [ 754.665520] env[62000]: value = "task-882087" [ 754.665520] env[62000]: _type = "Task" [ 754.665520] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.672088] env[62000]: INFO nova.compute.manager [-] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Took 1.46 seconds to deallocate network for instance. [ 754.680614] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882087, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.755035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a6731694-dbe3-44b4-b7aa-357bb500ae98 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.870s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.760417] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882081, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.814172] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.814503] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.941276] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d638f5-1693-c228-f5fd-4bfe9d5a2c88, 'name': SearchDatastore_Task, 'duration_secs': 0.031866} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.941276] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.941513] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 754.941764] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.941956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.942187] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 754.942459] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fe201c42-8a4e-4427-b72f-cfcba1163d74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.954783] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 754.954949] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 754.955707] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-17320196-43e9-4de0-b308-b9e1d215604a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.966123] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 754.966123] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526eb21a-bdea-f8ab-017b-d837b7a7e5cc" [ 754.966123] env[62000]: _type = "Task" [ 754.966123] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.974214] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526eb21a-bdea-f8ab-017b-d837b7a7e5cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.116202] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.152282] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb9e51b0-37eb-4a91-a22c-b24c8d6350d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.161808] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dbb5c38-eac3-419f-a851-3a3d7868abd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.199465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.200931] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1b9aac-701b-429b-bf04-0dbd92139f10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.210489] env[62000]: DEBUG oslo_vmware.api [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Task: {'id': task-882087, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147504} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.211192] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 755.212910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 755.212910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 755.212910] env[62000]: INFO nova.compute.manager [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Took 1.12 seconds to destroy the instance on the hypervisor. [ 755.212910] env[62000]: DEBUG oslo.service.loopingcall [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.212910] env[62000]: DEBUG nova.compute.manager [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.212910] env[62000]: DEBUG nova.network.neutron [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 755.220864] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c38d6967-2416-40bf-b3a1-33eeb6b41e5c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.234641] env[62000]: DEBUG nova.compute.provider_tree [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.238293] env[62000]: DEBUG nova.network.neutron [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.258048] env[62000]: DEBUG oslo_vmware.api [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882081, 'name': PowerOnVM_Task, 'duration_secs': 1.419363} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.258342] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 755.258546] env[62000]: INFO nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Took 9.33 seconds to spawn the instance on the hypervisor. [ 755.258730] env[62000]: DEBUG nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 755.260538] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13691e72-f6ca-48df-975a-90cdc52834af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.321517] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.478873] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526eb21a-bdea-f8ab-017b-d837b7a7e5cc, 'name': SearchDatastore_Task, 'duration_secs': 0.010168} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.479690] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a3587b6-37eb-435e-a072-6d8a51180c6e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.485906] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 755.485906] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527af230-f6b6-4e58-2c35-0affb4dbf97f" [ 755.485906] env[62000]: _type = "Task" [ 755.485906] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.495487] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527af230-f6b6-4e58-2c35-0affb4dbf97f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.739347] env[62000]: DEBUG nova.scheduler.client.report [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 755.742014] env[62000]: DEBUG nova.network.neutron [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.783853] env[62000]: INFO nova.compute.manager [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Took 31.52 seconds to build instance. [ 755.794565] env[62000]: DEBUG nova.compute.manager [req-30064944-b18b-40b7-80f5-8829eb63b0a0 req-8d7b806e-3ec2-4994-8bda-67e649f5a8d7 service nova] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Received event network-vif-deleted-1fd8ccbc-54eb-43e8-9f25-cadbff2a3fa8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 755.847818] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 755.998045] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527af230-f6b6-4e58-2c35-0affb4dbf97f, 'name': SearchDatastore_Task, 'duration_secs': 0.028166} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.998288] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.998562] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 30d82c70-1401-4a1a-a88b-f798f8fbf96a/30d82c70-1401-4a1a-a88b-f798f8fbf96a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 755.998849] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-288da3f0-c4cf-46d7-9dbd-2d3b3921e9d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.003053] env[62000]: DEBUG nova.compute.manager [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-changed-7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 756.003260] env[62000]: DEBUG nova.compute.manager [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing instance network info cache due to event network-changed-7961bfcb-4d94-4279-bc3a-b46ca382b0cf. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 756.004270] env[62000]: DEBUG oslo_concurrency.lockutils [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.004270] env[62000]: DEBUG oslo_concurrency.lockutils [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.004270] env[62000]: DEBUG nova.network.neutron [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing network info cache for port 7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 756.013064] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 756.013064] env[62000]: value = "task-882088" [ 756.013064] env[62000]: _type = "Task" [ 756.013064] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.024206] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.199990] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.203029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.203029] env[62000]: INFO nova.compute.manager [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Attaching volume 4d5a6ee5-2bf9-4271-9d97-7f85e831f836 to /dev/sdb [ 756.248183] env[62000]: INFO nova.compute.manager [-] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Took 1.03 seconds to deallocate network for instance. [ 756.248183] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.455s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.248183] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.253852] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.023s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.256417] env[62000]: INFO nova.compute.claims [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.260209] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-304a1e96-f93f-47a7-8de1-5db955c38117 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.277444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d487cd08-d77e-4117-bafa-ae53adbe8cff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.287676] env[62000]: DEBUG oslo_concurrency.lockutils [None req-874abe92-c1b4-45a3-9068-35e9b420065c tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.328s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.294188] env[62000]: DEBUG nova.virt.block_device [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating existing volume attachment record: 57e11cc3-168e-4262-be19-aa134b61edcb {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 756.528709] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882088, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.771734] env[62000]: DEBUG nova.compute.utils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 756.776217] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.777096] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 756.777697] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 757.027169] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.550009} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 757.027499] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 30d82c70-1401-4a1a-a88b-f798f8fbf96a/30d82c70-1401-4a1a-a88b-f798f8fbf96a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 757.027729] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 757.028025] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ff0c8daf-8e43-41da-b34e-ac8394d9367e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.036047] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 757.036047] env[62000]: value = "task-882092" [ 757.036047] env[62000]: _type = "Task" [ 757.036047] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 757.046681] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882092, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.091877] env[62000]: DEBUG nova.policy [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '83fc5232a1484305b2800d56f539c251', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0b5041eab4c4480c9221aefa7838ab27', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.164542] env[62000]: DEBUG nova.network.neutron [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updated VIF entry in instance network info cache for port 7961bfcb-4d94-4279-bc3a-b46ca382b0cf. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 757.164939] env[62000]: DEBUG nova.network.neutron [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.277080] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 757.515953] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Successfully created port: c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 757.552500] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882092, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 757.650464] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cfd8bba-0939-4d29-8787-ea5aac6d43e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.662685] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25abde57-4c92-4919-ab84-4eb9b381224c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.667363] env[62000]: DEBUG oslo_concurrency.lockutils [req-03d69375-1f48-4bd4-8953-99de6b6d431a req-b4c6eb82-2ed1-4f27-af85-9385a4b4fda1 service nova] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.698748] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1872f58b-8ab7-47d8-9c8e-5c38755eadbe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.707870] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-943fc7f6-ee7e-45e2-be37-96bdefde6386 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.725768] env[62000]: DEBUG nova.compute.provider_tree [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.047702] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882092, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.529973} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 758.047960] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 758.049099] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0148c348-1652-4a23-8869-123b7af47efb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.074622] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Reconfiguring VM instance instance-00000033 to attach disk [datastore2] 30d82c70-1401-4a1a-a88b-f798f8fbf96a/30d82c70-1401-4a1a-a88b-f798f8fbf96a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 758.076952] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc96186d-6bb6-4dd2-bd9e-c644b03a7b4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.109362] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 758.109362] env[62000]: value = "task-882093" [ 758.109362] env[62000]: _type = "Task" [ 758.109362] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 758.120123] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882093, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.228970] env[62000]: DEBUG nova.scheduler.client.report [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 758.300532] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 758.330009] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 758.330369] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 758.330661] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 758.330993] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 758.331292] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 758.331561] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 758.331899] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 758.332169] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 758.332375] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 758.332554] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 758.332737] env[62000]: DEBUG nova.virt.hardware [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 758.333637] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ec8465-6d54-46f6-b0a0-1b823982fafd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.342740] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03a1502-67f3-47cc-8106-b994c5bb9822 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.606163] env[62000]: DEBUG nova.compute.manager [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 758.622607] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882093, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 758.735686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.482s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.736172] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 758.738749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.205s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.740130] env[62000]: INFO nova.compute.claims [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 759.078144] env[62000]: DEBUG nova.compute.manager [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-vif-plugged-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 759.078144] env[62000]: DEBUG oslo_concurrency.lockutils [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] Acquiring lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.078144] env[62000]: DEBUG oslo_concurrency.lockutils [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.078144] env[62000]: DEBUG oslo_concurrency.lockutils [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.078144] env[62000]: DEBUG nova.compute.manager [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] No waiting events found dispatching network-vif-plugged-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 759.078144] env[62000]: WARNING nova.compute.manager [req-2c037131-e013-44b5-9a15-9b174ac1ea90 req-58981bd1-b4a6-4f7f-a29e-7c3b2d92a4b2 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received unexpected event network-vif-plugged-c4e0acd0-a5a2-4120-86ff-826f7dbe160e for instance with vm_state building and task_state spawning. [ 759.119963] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.120394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.120631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.120822] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.120994] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.123770] env[62000]: INFO nova.compute.manager [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Terminating instance [ 759.128993] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882093, 'name': ReconfigVM_Task, 'duration_secs': 0.744955} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.131067] env[62000]: DEBUG nova.compute.manager [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.131067] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 759.131275] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Reconfigured VM instance instance-00000033 to attach disk [datastore2] 30d82c70-1401-4a1a-a88b-f798f8fbf96a/30d82c70-1401-4a1a-a88b-f798f8fbf96a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 759.132837] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4746598-764b-461e-a2bc-ed110e3ea9ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.138426] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.138743] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-df1928a7-4a1b-437f-a2ea-5bd4b4a29532 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.147373] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 759.147642] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c89c3b1-f6fa-46c3-8bef-fbecc01874ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.151765] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 759.151765] env[62000]: value = "task-882095" [ 759.151765] env[62000]: _type = "Task" [ 759.151765] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.158207] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 759.158207] env[62000]: value = "task-882096" [ 759.158207] env[62000]: _type = "Task" [ 759.158207] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.161352] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882095, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.170293] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-882096, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.207733] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.208078] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.220623] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Successfully updated port: c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 759.246490] env[62000]: DEBUG nova.compute.utils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.248301] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.248509] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 759.320907] env[62000]: DEBUG nova.policy [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cae2549ebc1d4403bf7148d6b77b3b27', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3982d7c213344f7d822c2adb0ad36bb4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 759.670796] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882095, 'name': Rename_Task, 'duration_secs': 0.256828} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.671606] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 759.671960] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8535d06c-7bba-4d4c-8d19-7270e2cb9f23 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.679168] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-882096, 'name': PowerOffVM_Task, 'duration_secs': 0.31874} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 759.679814] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Successfully created port: 0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 759.686022] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 759.686022] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 759.686843] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-99764487-b694-4bc4-8829-6d8a7340266a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.689392] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 759.689392] env[62000]: value = "task-882097" [ 759.689392] env[62000]: _type = "Task" [ 759.689392] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.698251] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882097, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.710894] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.723179] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.723453] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.723589] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 759.752680] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 759.763989] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 759.765013] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 759.765013] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Deleting the datastore file [datastore1] 49f6c198-13b7-4c07-81d8-c010e7b0598e {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 759.765013] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-07ee0844-21b7-4082-ae75-b82669baabbc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.777582] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for the task: (returnval){ [ 759.777582] env[62000]: value = "task-882099" [ 759.777582] env[62000]: _type = "Task" [ 759.777582] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.795903] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-882099, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.031414] env[62000]: DEBUG nova.compute.manager [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.032363] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb9fd44-7bef-4d0a-abf9-cec6163ce201 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.153524] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f61a960-f043-440e-8e3d-152eeee34213 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.161627] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71563884-cf7b-420a-ba94-96e85ebe3f8f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.196501] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3522a17f-d0f4-46c8-a803-da827bd80ed1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.207194] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882097, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.208410] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4410d810-4bbd-4572-b20e-30182ed8b8be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.227534] env[62000]: DEBUG nova.compute.provider_tree [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.241069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 760.262146] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.291031] env[62000]: DEBUG oslo_vmware.api [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Task: {'id': task-882099, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.244158} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.291031] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.291031] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 760.291031] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 760.291330] env[62000]: INFO nova.compute.manager [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Took 1.16 seconds to destroy the instance on the hypervisor. [ 760.291430] env[62000]: DEBUG oslo.service.loopingcall [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.291632] env[62000]: DEBUG nova.compute.manager [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.291787] env[62000]: DEBUG nova.network.neutron [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 760.544103] env[62000]: DEBUG nova.network.neutron [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.548147] env[62000]: INFO nova.compute.manager [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] instance snapshotting [ 760.548147] env[62000]: DEBUG nova.objects.instance [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.703444] env[62000]: DEBUG oslo_vmware.api [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882097, 'name': PowerOnVM_Task, 'duration_secs': 0.712728} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.703444] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 760.703621] env[62000]: INFO nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Took 6.93 seconds to spawn the instance on the hypervisor. [ 760.703752] env[62000]: DEBUG nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 760.704557] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a92914-4506-42e5-b3e9-68fac113f6fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.732589] env[62000]: DEBUG nova.scheduler.client.report [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.765688] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 760.789968] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 760.790253] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 760.790417] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 760.790606] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 760.791174] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 760.791174] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 760.791174] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 760.791354] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 760.791463] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 760.791634] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 760.791831] env[62000]: DEBUG nova.virt.hardware [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 760.792835] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623979b1-d3a7-4067-9dbb-ce16af9e6649 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.801687] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49ea5fd4-bca8-4718-ab92-50e42475f4e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.049439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.049787] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Instance network_info: |[{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 761.050244] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e0:5d:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '674802e7-b847-4bef-a7a8-f90ac7a3a0a7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c4e0acd0-a5a2-4120-86ff-826f7dbe160e', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.057663] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Creating folder: Project (0b5041eab4c4480c9221aefa7838ab27). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 761.060546] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-01d48931-dd31-4a16-adac-fd9d983b01ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.063205] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da6f861b-493a-4fc2-a108-d8c8da0fab2e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.083926] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247db94b-6fc7-4798-b126-77ba6f9fe7f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.086992] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Created folder: Project (0b5041eab4c4480c9221aefa7838ab27) in parent group-v201431. [ 761.086992] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Creating folder: Instances. Parent ref: group-v201511. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 761.087202] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0ddd84dc-672e-46cd-a5d4-7a3f764f017c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.099208] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Created folder: Instances in parent group-v201511. [ 761.099470] env[62000]: DEBUG oslo.service.loopingcall [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.099670] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 761.099879] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-014020d8-3dbe-40d8-b702-e8266cf3a108 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.116606] env[62000]: DEBUG nova.compute.manager [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 761.116810] env[62000]: DEBUG nova.compute.manager [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing instance network info cache due to event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 761.117032] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.117191] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.117358] env[62000]: DEBUG nova.network.neutron [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 761.125274] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.125274] env[62000]: value = "task-882102" [ 761.125274] env[62000]: _type = "Task" [ 761.125274] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.135681] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882102, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.222451] env[62000]: INFO nova.compute.manager [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Took 33.61 seconds to build instance. [ 761.237647] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.238248] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 761.241422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.410s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.241595] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.244781] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 761.244781] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.365s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.244781] env[62000]: INFO nova.compute.claims [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.246977] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12204fdf-7dbf-45c8-b150-e95a4f52b32b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.257388] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b56d32-b880-473d-9ee6-e8c9726e88d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.274879] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cde96670-a8b7-44f1-bd32-9f30d08bdf4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.284614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ea9df3c-2d66-4e50-8d41-5ab4369f5766 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.318988] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180908MB free_disk=52GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 761.319724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.369477] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 761.369736] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201510', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'name': 'volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'eb27703f-b657-423a-90a9-a7c024a2e473', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'serial': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 761.370815] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1638055-e2da-422d-bec6-d65ac92a8685 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.389561] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec441c2e-9e9c-4664-a7e9-bdb9aecce85f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.415332] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfiguring VM instance instance-00000024 to attach disk [datastore2] volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836/volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 761.415652] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-311cd8a9-5271-4363-98b9-802514186ef0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.436102] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 761.436102] env[62000]: value = "task-882103" [ 761.436102] env[62000]: _type = "Task" [ 761.436102] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.445156] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882103, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.509607] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Successfully updated port: 0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 761.528771] env[62000]: DEBUG nova.network.neutron [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.598755] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 761.599153] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-48d57547-24d1-4030-a2e9-208dd3c9a6c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.613863] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 761.613863] env[62000]: value = "task-882104" [ 761.613863] env[62000]: _type = "Task" [ 761.613863] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.626253] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882104, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.635747] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882102, 'name': CreateVM_Task, 'duration_secs': 0.489916} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.635877] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 761.636607] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.636860] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.637298] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.637588] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4675d0c-5b5e-46d2-b3fd-dd3256710a4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.643877] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 761.643877] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52989df9-e551-046a-6a1d-45feb374b7e5" [ 761.643877] env[62000]: _type = "Task" [ 761.643877] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.655623] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52989df9-e551-046a-6a1d-45feb374b7e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.724623] env[62000]: DEBUG oslo_concurrency.lockutils [None req-46eea6e6-3b50-40c9-bdfa-8ae40a89c5b2 tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 97.669s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.750560] env[62000]: DEBUG nova.compute.utils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 761.752541] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 761.752756] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 761.854012] env[62000]: DEBUG nova.policy [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '30019e05995d4da88096c54fc116e9c1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3b0afa95373a4322b3b14093516a8e49', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 761.948129] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882103, 'name': ReconfigVM_Task, 'duration_secs': 0.40465} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.948492] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfigured VM instance instance-00000024 to attach disk [datastore2] volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836/volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 761.960237] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0288a7e7-e9e2-4fe5-abe4-812d53420814 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.979038] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 761.979038] env[62000]: value = "task-882105" [ 761.979038] env[62000]: _type = "Task" [ 761.979038] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.989725] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.016211] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.016387] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquired lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.016564] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.034419] env[62000]: INFO nova.compute.manager [-] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Took 1.74 seconds to deallocate network for instance. [ 762.066764] env[62000]: DEBUG nova.network.neutron [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updated VIF entry in instance network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 762.067053] env[62000]: DEBUG nova.network.neutron [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.125439] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882104, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.155579] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52989df9-e551-046a-6a1d-45feb374b7e5, 'name': SearchDatastore_Task, 'duration_secs': 0.021391} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.155917] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.156234] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.156481] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.156631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.156817] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.157110] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ef1bf745-a104-40ee-9802-4f4f40eb4cc6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.168552] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.168770] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 762.169710] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bbd42b89-7a2d-4db8-962a-5ddea92833cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.175692] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 762.175692] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c18a62-f247-f1a9-99df-69a48ae0e692" [ 762.175692] env[62000]: _type = "Task" [ 762.175692] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.185612] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c18a62-f247-f1a9-99df-69a48ae0e692, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.212163] env[62000]: DEBUG nova.compute.manager [None req-ff7eb594-f6c7-4cc3-be39-00908efe139c tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 762.213147] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4be0098-15d4-416d-b0cc-01c33f7ac13e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.257744] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 762.284967] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Successfully created port: b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 762.434171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.434498] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.434795] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.435060] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.435272] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.437853] env[62000]: INFO nova.compute.manager [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Terminating instance [ 762.439918] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "refresh_cache-30d82c70-1401-4a1a-a88b-f798f8fbf96a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.440200] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquired lock "refresh_cache-30d82c70-1401-4a1a-a88b-f798f8fbf96a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.440283] env[62000]: DEBUG nova.network.neutron [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 762.489772] env[62000]: DEBUG oslo_vmware.api [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882105, 'name': ReconfigVM_Task, 'duration_secs': 0.164602} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.490448] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201510', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'name': 'volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'eb27703f-b657-423a-90a9-a7c024a2e473', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'serial': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 762.541261] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.563365] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.570440] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.570715] env[62000]: DEBUG nova.compute.manager [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Received event network-vif-deleted-c0059cab-c384-49ca-b0fa-1e46163e688c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 762.571069] env[62000]: INFO nova.compute.manager [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Neutron deleted interface c0059cab-c384-49ca-b0fa-1e46163e688c; detaching it from the instance and deleting it from the info cache [ 762.571125] env[62000]: DEBUG nova.network.neutron [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.632471] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882104, 'name': CreateSnapshot_Task, 'duration_secs': 0.614142} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.632845] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 762.634135] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc0f37b4-75c5-4b6e-bb1f-126175707a4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.653306] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46a3bdc6-f584-48b1-89c8-bccffb76e550 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.664369] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b40cbd25-6486-4e34-ae6d-d00cd340853e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.701449] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4737d841-c537-43d5-9682-a10552a85003 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.713696] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f996e4-1cae-4fc0-b382-116e18c1cf43 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.717834] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c18a62-f247-f1a9-99df-69a48ae0e692, 'name': SearchDatastore_Task, 'duration_secs': 0.01846} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.718963] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55655ace-48c8-4929-ad05-ee027a4ac28a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.731779] env[62000]: INFO nova.compute.manager [None req-ff7eb594-f6c7-4cc3-be39-00908efe139c tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] instance snapshotting [ 762.732448] env[62000]: DEBUG nova.objects.instance [None req-ff7eb594-f6c7-4cc3-be39-00908efe139c tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lazy-loading 'flavor' on Instance uuid 30d82c70-1401-4a1a-a88b-f798f8fbf96a {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 762.734707] env[62000]: DEBUG nova.compute.provider_tree [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.739495] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 762.739495] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52382f6e-024f-c6b1-ad70-2d8f9e245621" [ 762.739495] env[62000]: _type = "Task" [ 762.739495] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.739495] env[62000]: DEBUG nova.network.neutron [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Updating instance_info_cache with network_info: [{"id": "0c52b803-0943-434c-a133-1f05adb65a75", "address": "fa:16:3e:ae:f2:bc", "network": {"id": "345ecafa-d306-4609-bcce-075d6c501bd9", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1348686142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3982d7c213344f7d822c2adb0ad36bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c52b803-09", "ovs_interfaceid": "0c52b803-0943-434c-a133-1f05adb65a75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.750082] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52382f6e-024f-c6b1-ad70-2d8f9e245621, 'name': SearchDatastore_Task, 'duration_secs': 0.020594} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.750082] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.750082] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] ea8ae421-1d7f-4814-bc0f-90a3316ad028/ea8ae421-1d7f-4814-bc0f-90a3316ad028.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 762.750304] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-919f62c6-e5d6-48fa-bd78-b0d9b108e621 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.762388] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 762.762388] env[62000]: value = "task-882106" [ 762.762388] env[62000]: _type = "Task" [ 762.762388] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.774679] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882106, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.958871] env[62000]: DEBUG nova.network.neutron [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 763.017325] env[62000]: DEBUG nova.network.neutron [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.077858] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1278fa72-0cb6-4da8-aaf3-57a2e222f161 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.093195] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ff4a77-dfb0-4582-ad91-47c06564662d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.126699] env[62000]: DEBUG nova.compute.manager [req-a5b0500a-2e1f-4526-bf5c-e4aef6a78bd8 req-2c021bd7-288a-463d-b49a-ba868095c51c service nova] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Detach interface failed, port_id=c0059cab-c384-49ca-b0fa-1e46163e688c, reason: Instance 49f6c198-13b7-4c07-81d8-c010e7b0598e could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 763.132826] env[62000]: DEBUG nova.compute.manager [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Received event network-vif-plugged-0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.132826] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Acquiring lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.133189] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.133372] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.133413] env[62000]: DEBUG nova.compute.manager [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] No waiting events found dispatching network-vif-plugged-0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 763.133554] env[62000]: WARNING nova.compute.manager [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Received unexpected event network-vif-plugged-0c52b803-0943-434c-a133-1f05adb65a75 for instance with vm_state building and task_state spawning. [ 763.133719] env[62000]: DEBUG nova.compute.manager [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Received event network-changed-0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 763.133876] env[62000]: DEBUG nova.compute.manager [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Refreshing instance network info cache due to event network-changed-0c52b803-0943-434c-a133-1f05adb65a75. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 763.134411] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Acquiring lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.164209] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 763.164555] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d6da1c17-f6a5-4343-a125-ac19a6284df5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.178461] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 763.178461] env[62000]: value = "task-882107" [ 763.178461] env[62000]: _type = "Task" [ 763.178461] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.189229] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882107, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.244764] env[62000]: DEBUG nova.scheduler.client.report [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.246713] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Releasing lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.247285] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Instance network_info: |[{"id": "0c52b803-0943-434c-a133-1f05adb65a75", "address": "fa:16:3e:ae:f2:bc", "network": {"id": "345ecafa-d306-4609-bcce-075d6c501bd9", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1348686142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3982d7c213344f7d822c2adb0ad36bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c52b803-09", "ovs_interfaceid": "0c52b803-0943-434c-a133-1f05adb65a75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 763.248414] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-689cfbe7-a6d3-47e9-9ed5-918d3d0143fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.253903] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Acquired lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.256491] env[62000]: DEBUG nova.network.neutron [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Refreshing network info cache for port 0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 763.259602] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:f2:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '78e1ebb0-0130-446b-bf73-a0e59bbb95cc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c52b803-0943-434c-a133-1f05adb65a75', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 763.273407] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Creating folder: Project (3982d7c213344f7d822c2adb0ad36bb4). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.277035] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 763.280495] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3894b309-9886-4b5f-abb5-c48558139c43 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.311414] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-067f62e7-72d2-4200-bc32-f7c689c3b27f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.319027] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882106, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.321156] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Created folder: Project (3982d7c213344f7d822c2adb0ad36bb4) in parent group-v201431. [ 763.321465] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Creating folder: Instances. Parent ref: group-v201516. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 763.321835] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d712e4bf-e8de-4402-b2d3-865ab3757e7d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 763.328258] env[62000]: DEBUG nova.virt.hardware [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 763.329088] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34d0916c-c23f-463a-bbca-d99030220b95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.345276] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ed39c84-5a74-47a5-a25e-7f2e705a607a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.351629] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Created folder: Instances in parent group-v201516. [ 763.351885] env[62000]: DEBUG oslo.service.loopingcall [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 763.352481] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 763.353260] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9730441-86c3-4ff5-be04-26b4c9b15727 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.384352] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 763.384352] env[62000]: value = "task-882110" [ 763.384352] env[62000]: _type = "Task" [ 763.384352] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.392591] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882110, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.520681] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Releasing lock "refresh_cache-30d82c70-1401-4a1a-a88b-f798f8fbf96a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.521166] env[62000]: DEBUG nova.compute.manager [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 763.521395] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 763.522367] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a814c453-537b-4fcd-82bc-695da8b708f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.530744] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 763.531023] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-344fe3c7-bc79-4362-b699-98cbe62019a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.540100] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 763.540100] env[62000]: value = "task-882111" [ 763.540100] env[62000]: _type = "Task" [ 763.540100] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.544440] env[62000]: DEBUG nova.objects.instance [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lazy-loading 'flavor' on Instance uuid eb27703f-b657-423a-90a9-a7c024a2e473 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.551101] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882111, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.690318] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882107, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.747117] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.505s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.748380] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.486s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.748612] env[62000]: DEBUG nova.objects.instance [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lazy-loading 'resources' on Instance uuid 1298037f-9cb7-4e2b-b70f-2ab24efe9b91 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 763.787978] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882106, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567343} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.788470] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] ea8ae421-1d7f-4814-bc0f-90a3316ad028/ea8ae421-1d7f-4814-bc0f-90a3316ad028.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 763.788723] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 763.788995] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11590735-ae7c-4dac-afac-c73a782fc098 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.797957] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 763.797957] env[62000]: value = "task-882112" [ 763.797957] env[62000]: _type = "Task" [ 763.797957] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.808757] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882112, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.839767] env[62000]: DEBUG nova.compute.manager [None req-ff7eb594-f6c7-4cc3-be39-00908efe139c tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance disappeared during snapshot {{(pid=62000) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 763.899105] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882110, 'name': CreateVM_Task, 'duration_secs': 0.407279} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.899105] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 763.899105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 763.899105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 763.899385] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 763.899534] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-959c0501-5482-4547-aafb-52c7e4bc7e19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.905350] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 763.905350] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f841e7-2e85-0b11-e632-ee738d88436b" [ 763.905350] env[62000]: _type = "Task" [ 763.905350] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.915022] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f841e7-2e85-0b11-e632-ee738d88436b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.991759] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Successfully updated port: b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 763.993721] env[62000]: DEBUG nova.compute.manager [None req-ff7eb594-f6c7-4cc3-be39-00908efe139c tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Found 0 images (rotation: 2) {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 764.058257] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882111, 'name': PowerOffVM_Task, 'duration_secs': 0.446631} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.058531] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 764.058736] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 764.060014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e92963ec-3d70-4db4-8061-d8788bad19c6 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.859s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.061277] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8cd4bc71-de36-49f1-975a-40878a00dbdc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.091548] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 764.091822] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 764.092112] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Deleting the datastore file [datastore2] 30d82c70-1401-4a1a-a88b-f798f8fbf96a {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 764.092367] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a7184c9f-e3b3-4a5b-bc72-0e6381a746f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.100504] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for the task: (returnval){ [ 764.100504] env[62000]: value = "task-882114" [ 764.100504] env[62000]: _type = "Task" [ 764.100504] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.115069] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882114, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.157130] env[62000]: DEBUG nova.network.neutron [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Updated VIF entry in instance network info cache for port 0c52b803-0943-434c-a133-1f05adb65a75. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 764.157543] env[62000]: DEBUG nova.network.neutron [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Updating instance_info_cache with network_info: [{"id": "0c52b803-0943-434c-a133-1f05adb65a75", "address": "fa:16:3e:ae:f2:bc", "network": {"id": "345ecafa-d306-4609-bcce-075d6c501bd9", "bridge": "br-int", "label": "tempest-InstanceActionsNegativeTestJSON-1348686142-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3982d7c213344f7d822c2adb0ad36bb4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "78e1ebb0-0130-446b-bf73-a0e59bbb95cc", "external-id": "nsx-vlan-transportzone-414", "segmentation_id": 414, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c52b803-09", "ovs_interfaceid": "0c52b803-0943-434c-a133-1f05adb65a75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.190647] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882107, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.251029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "fae7fa75-f8a9-4b9e-b878-852eb9d17b58" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.251162] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "fae7fa75-f8a9-4b9e-b878-852eb9d17b58" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.311032] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882112, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.081655} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.311337] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.312175] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bf365b3-7b86-4bc3-ac14-c437912abaa3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.337394] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] ea8ae421-1d7f-4814-bc0f-90a3316ad028/ea8ae421-1d7f-4814-bc0f-90a3316ad028.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.340259] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cac61944-6bb8-497d-b039-da747c56bb33 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.362620] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 764.362620] env[62000]: value = "task-882115" [ 764.362620] env[62000]: _type = "Task" [ 764.362620] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.377709] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882115, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.419043] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f841e7-2e85-0b11-e632-ee738d88436b, 'name': SearchDatastore_Task, 'duration_secs': 0.036816} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.422185] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.422560] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 764.422825] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.423018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.423263] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 764.423771] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b26f2d0b-f0b8-4588-8cd0-28517280fba2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.437101] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 764.437342] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 764.438108] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c6d8aa90-4034-40d0-b200-e611cfe997b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.444254] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 764.444254] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf4c9a-8fda-dba9-cf2e-f59532c100cb" [ 764.444254] env[62000]: _type = "Task" [ 764.444254] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.456616] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf4c9a-8fda-dba9-cf2e-f59532c100cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.496237] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.496385] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquired lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.496537] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.612922] env[62000]: DEBUG oslo_vmware.api [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Task: {'id': task-882114, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144453} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.615494] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 764.615696] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 764.615875] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 764.616072] env[62000]: INFO nova.compute.manager [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Took 1.09 seconds to destroy the instance on the hypervisor. [ 764.616334] env[62000]: DEBUG oslo.service.loopingcall [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 764.616763] env[62000]: DEBUG nova.compute.manager [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.616864] env[62000]: DEBUG nova.network.neutron [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 764.639623] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62a1872-cc9b-45e3-a0cd-4f8897ffe3b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.642760] env[62000]: DEBUG nova.network.neutron [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 764.650789] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d608cba0-ab2f-414e-b6c6-9662ca27f69a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.682021] env[62000]: DEBUG oslo_concurrency.lockutils [req-85a77a19-8052-46ec-b512-3ddb1748630e req-2e24c646-d319-4f99-8a00-37931d38819b service nova] Releasing lock "refresh_cache-df525129-0ccb-4863-8a22-dd3e5a1aa2b5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.685986] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bddb7d4b-d758-4b6b-9dcf-e02864db5a99 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.694342] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882107, 'name': CloneVM_Task, 'duration_secs': 1.452167} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.696420] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created linked-clone VM from snapshot [ 764.697471] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3130708e-d3b1-4906-aa10-4537bba65ce6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.701080] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8582c7be-dfe2-42f0-91d0-d23210b9ae52 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.715762] env[62000]: DEBUG nova.compute.provider_tree [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 764.719466] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploading image 150ffec2-eb6a-4b9b-baf4-6005539ac9fa {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 764.745198] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 764.745198] env[62000]: value = "vm-201515" [ 764.745198] env[62000]: _type = "VirtualMachine" [ 764.745198] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 764.745481] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-558894b7-32a7-4198-9cd4-6bd8a5b7dbab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.754335] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease: (returnval){ [ 764.754335] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52aefb54-0f76-ed27-8202-634e817b1190" [ 764.754335] env[62000]: _type = "HttpNfcLease" [ 764.754335] env[62000]: } obtained for exporting VM: (result){ [ 764.754335] env[62000]: value = "vm-201515" [ 764.754335] env[62000]: _type = "VirtualMachine" [ 764.754335] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 764.754661] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the lease: (returnval){ [ 764.754661] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52aefb54-0f76-ed27-8202-634e817b1190" [ 764.754661] env[62000]: _type = "HttpNfcLease" [ 764.754661] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 764.758358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "fae7fa75-f8a9-4b9e-b878-852eb9d17b58" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.507s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.758877] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 764.765779] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 764.765779] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52aefb54-0f76-ed27-8202-634e817b1190" [ 764.765779] env[62000]: _type = "HttpNfcLease" [ 764.765779] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 764.873850] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882115, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.954627] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf4c9a-8fda-dba9-cf2e-f59532c100cb, 'name': SearchDatastore_Task, 'duration_secs': 0.013082} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.955406] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8ed14a1-f96f-43ed-a39d-4d9c23146869 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.961354] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 764.961354] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52826ebd-1f29-7284-15c7-35c81fe6d436" [ 764.961354] env[62000]: _type = "Task" [ 764.961354] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.969202] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52826ebd-1f29-7284-15c7-35c81fe6d436, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.977702] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.977974] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.042487] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.146264] env[62000]: DEBUG nova.network.neutron [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.195625] env[62000]: DEBUG nova.compute.manager [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Received event network-vif-plugged-b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.195872] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Acquiring lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.196102] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.196290] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.196461] env[62000]: DEBUG nova.compute.manager [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] No waiting events found dispatching network-vif-plugged-b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 765.196630] env[62000]: WARNING nova.compute.manager [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Received unexpected event network-vif-plugged-b6da36aa-e361-46b2-99a3-7c942b5ae516 for instance with vm_state building and task_state spawning. [ 765.196794] env[62000]: DEBUG nova.compute.manager [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Received event network-changed-b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 765.196951] env[62000]: DEBUG nova.compute.manager [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Refreshing instance network info cache due to event network-changed-b6da36aa-e361-46b2-99a3-7c942b5ae516. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 765.197342] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Acquiring lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 765.222616] env[62000]: DEBUG nova.scheduler.client.report [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.230370] env[62000]: DEBUG nova.network.neutron [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Updating instance_info_cache with network_info: [{"id": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "address": "fa:16:3e:23:a5:29", "network": {"id": "f8c30571-3229-4552-b3f8-ba4856e9d6a3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1424940862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b0afa95373a4322b3b14093516a8e49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6da36aa-e3", "ovs_interfaceid": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.265068] env[62000]: DEBUG nova.compute.utils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 765.266612] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 765.266612] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52aefb54-0f76-ed27-8202-634e817b1190" [ 765.266612] env[62000]: _type = "HttpNfcLease" [ 765.266612] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 765.267158] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 765.267460] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 765.269138] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 765.269138] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52aefb54-0f76-ed27-8202-634e817b1190" [ 765.269138] env[62000]: _type = "HttpNfcLease" [ 765.269138] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 765.269883] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bd562e-9709-42b1-a28b-8dc381690dac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.279479] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 765.279667] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 765.338240] env[62000]: DEBUG nova.policy [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a1f50fc5914d4f4ba54b541fc61924a1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4db75d84722b4f5faf90c531ffee3865', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 765.369834] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-52ca1a99-3a9b-4d6a-b47e-f4e0c4973d6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.377315] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882115, 'name': ReconfigVM_Task, 'duration_secs': 0.77704} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.378520] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfigured VM instance instance-00000034 to attach disk [datastore1] ea8ae421-1d7f-4814-bc0f-90a3316ad028/ea8ae421-1d7f-4814-bc0f-90a3316ad028.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 765.380836] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0edbde8b-ee82-4f2c-91a9-f8b52f7f595a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.390246] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 765.390246] env[62000]: value = "task-882117" [ 765.390246] env[62000]: _type = "Task" [ 765.390246] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.403393] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882117, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.472826] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52826ebd-1f29-7284-15c7-35c81fe6d436, 'name': SearchDatastore_Task, 'duration_secs': 0.009586} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.473124] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.473395] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] df525129-0ccb-4863-8a22-dd3e5a1aa2b5/df525129-0ccb-4863-8a22-dd3e5a1aa2b5.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 765.473658] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2951a966-edc4-41f9-89f2-156e9422177d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.480519] env[62000]: INFO nova.compute.manager [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Detaching volume 4d5a6ee5-2bf9-4271-9d97-7f85e831f836 [ 765.483847] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 765.483847] env[62000]: value = "task-882118" [ 765.483847] env[62000]: _type = "Task" [ 765.483847] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.497239] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882118, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.531150] env[62000]: INFO nova.virt.block_device [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Attempting to driver detach volume 4d5a6ee5-2bf9-4271-9d97-7f85e831f836 from mountpoint /dev/sdb [ 765.531565] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 765.531867] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201510', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'name': 'volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'eb27703f-b657-423a-90a9-a7c024a2e473', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'serial': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 765.532931] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feb13b3f-4d10-426f-a177-b5127b170cf1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.557314] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219fab98-a50c-4414-8e9b-1e74de52f25c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.566425] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5a0078-91df-495f-92dd-6c2a75fa6b20 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.589652] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e58ad6c-abd2-4221-bbb5-5c1183ebd1fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.606719] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] The volume has not been displaced from its original location: [datastore2] volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836/volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 765.615406] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfiguring VM instance instance-00000024 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 765.615406] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Successfully created port: c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 765.615679] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2dd31a1-2878-4da9-be13-eb8d9f340947 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.635648] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 765.635648] env[62000]: value = "task-882119" [ 765.635648] env[62000]: _type = "Task" [ 765.635648] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.645692] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882119, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.648875] env[62000]: INFO nova.compute.manager [-] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Took 1.03 seconds to deallocate network for instance. [ 765.730253] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.982s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.732884] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.269s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.735466] env[62000]: INFO nova.compute.claims [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 765.738668] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Releasing lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.738991] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Instance network_info: |[{"id": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "address": "fa:16:3e:23:a5:29", "network": {"id": "f8c30571-3229-4552-b3f8-ba4856e9d6a3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1424940862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b0afa95373a4322b3b14093516a8e49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6da36aa-e3", "ovs_interfaceid": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 765.739541] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Acquired lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 765.739828] env[62000]: DEBUG nova.network.neutron [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Refreshing network info cache for port b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 765.741058] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:a5:29', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f016d1-34a6-4ebd-81ed-a6bf9d109b87', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b6da36aa-e361-46b2-99a3-7c942b5ae516', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.752095] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Creating folder: Project (3b0afa95373a4322b3b14093516a8e49). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.752095] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b04eae7c-557f-4228-8d01-29ee6d6802d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.765518] env[62000]: INFO nova.scheduler.client.report [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Deleted allocations for instance 1298037f-9cb7-4e2b-b70f-2ab24efe9b91 [ 765.768605] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 765.777100] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Created folder: Project (3b0afa95373a4322b3b14093516a8e49) in parent group-v201431. [ 765.777385] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Creating folder: Instances. Parent ref: group-v201519. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 765.778212] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0c7d9933-0aca-4312-8bec-f8ba74cc56f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.794315] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Created folder: Instances in parent group-v201519. [ 765.796019] env[62000]: DEBUG oslo.service.loopingcall [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.796019] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 765.796019] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5289c05c-274c-4887-ad10-09edcf0e0f2f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.822661] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.822661] env[62000]: value = "task-882122" [ 765.822661] env[62000]: _type = "Task" [ 765.822661] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.838739] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882122, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.904367] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882117, 'name': Rename_Task, 'duration_secs': 0.140551} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.904367] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 765.904647] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a44f5d01-26ff-4c27-860b-88e5984b3a8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.913668] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 765.913668] env[62000]: value = "task-882123" [ 765.913668] env[62000]: _type = "Task" [ 765.913668] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.924059] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882123, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.998896] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882118, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502611} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 765.999446] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] df525129-0ccb-4863-8a22-dd3e5a1aa2b5/df525129-0ccb-4863-8a22-dd3e5a1aa2b5.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 765.999863] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 766.000485] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9ae8c562-e54b-40b1-a21d-bcc55c34d698 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.011823] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 766.011823] env[62000]: value = "task-882124" [ 766.011823] env[62000]: _type = "Task" [ 766.011823] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.025696] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882124, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.156071] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882119, 'name': ReconfigVM_Task, 'duration_secs': 0.31324} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.156134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.156488] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Reconfigured VM instance instance-00000024 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 766.161867] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7db242fb-eeda-4f8a-8353-426900d615a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.186087] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 766.186087] env[62000]: value = "task-882125" [ 766.186087] env[62000]: _type = "Task" [ 766.186087] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.197276] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882125, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.284905] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7359c8c-458d-4eb7-8a1e-a5212b4110d3 tempest-ImagesOneServerTestJSON-377593690 tempest-ImagesOneServerTestJSON-377593690-project-member] Lock "1298037f-9cb7-4e2b-b70f-2ab24efe9b91" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.395s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.335131] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882122, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.426331] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882123, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.523806] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882124, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073174} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.524189] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 766.525104] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c301584-1200-4fe2-a5aa-1cbf517145c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.549839] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Reconfiguring VM instance instance-00000035 to attach disk [datastore1] df525129-0ccb-4863-8a22-dd3e5a1aa2b5/df525129-0ccb-4863-8a22-dd3e5a1aa2b5.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 766.553350] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e76a83ab-de92-454e-95d3-98858ba9f3a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.575913] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 766.575913] env[62000]: value = "task-882126" [ 766.575913] env[62000]: _type = "Task" [ 766.575913] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.587091] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882126, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.705900] env[62000]: DEBUG oslo_vmware.api [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882125, 'name': ReconfigVM_Task, 'duration_secs': 0.179943} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.706744] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201510', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'name': 'volume-4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'eb27703f-b657-423a-90a9-a7c024a2e473', 'attached_at': '', 'detached_at': '', 'volume_id': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836', 'serial': '4d5a6ee5-2bf9-4271-9d97-7f85e831f836'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 766.753920] env[62000]: DEBUG nova.network.neutron [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Updated VIF entry in instance network info cache for port b6da36aa-e361-46b2-99a3-7c942b5ae516. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 766.754418] env[62000]: DEBUG nova.network.neutron [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Updating instance_info_cache with network_info: [{"id": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "address": "fa:16:3e:23:a5:29", "network": {"id": "f8c30571-3229-4552-b3f8-ba4856e9d6a3", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1424940862-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "3b0afa95373a4322b3b14093516a8e49", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f016d1-34a6-4ebd-81ed-a6bf9d109b87", "external-id": "nsx-vlan-transportzone-629", "segmentation_id": 629, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb6da36aa-e3", "ovs_interfaceid": "b6da36aa-e361-46b2-99a3-7c942b5ae516", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.788156] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 766.817812] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 766.818203] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 766.818445] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 766.818674] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 766.819034] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 766.819132] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 766.819691] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 766.819691] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 766.819691] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 766.820047] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 766.820108] env[62000]: DEBUG nova.virt.hardware [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 766.821057] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32c8363d-1530-4ff2-b68e-bbc53ad07574 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.840059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f66b4d-9608-40eb-a684-c53d2f2fa4ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.848451] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882122, 'name': CreateVM_Task, 'duration_secs': 0.581518} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.851622] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 766.853069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.853069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.857019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.857019] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00a84f37-860a-4202-8a90-0f068290c2d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.871229] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 766.871229] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52245834-68e9-a168-51f1-8a95d0037959" [ 766.871229] env[62000]: _type = "Task" [ 766.871229] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.884779] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52245834-68e9-a168-51f1-8a95d0037959, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.925367] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882123, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.090730] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882126, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.160626] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-618932d9-7f08-4f68-9743-368b292631a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.169690] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-678a7c64-98c1-4e28-8c51-8a9da36e36ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.204734] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b9dbf4c-c487-4da8-aac1-1fd48c038c2d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.214083] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a0720f-59c7-4bf2-985e-127107eeae47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.232415] env[62000]: DEBUG nova.compute.provider_tree [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.256817] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0d64f16-89c8-4ad7-a0a0-080c668e4077 req-1186c2fe-cf4c-494a-8216-b13761976534 service nova] Releasing lock "refresh_cache-4c49fdd0-2485-4791-9349-a79a8663bbc0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.294078] env[62000]: DEBUG nova.objects.instance [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lazy-loading 'flavor' on Instance uuid eb27703f-b657-423a-90a9-a7c024a2e473 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 767.384583] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52245834-68e9-a168-51f1-8a95d0037959, 'name': SearchDatastore_Task, 'duration_secs': 0.027095} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.385077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.385393] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 767.385785] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 767.385992] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 767.386323] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 767.386685] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-794f3034-ca8c-49b4-bd7c-b5912a2f8cdf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.398135] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 767.398689] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 767.399885] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-910d3c73-31b7-41e0-9473-8b984c83b73f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.407339] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 767.407339] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524307c4-2e1d-dee2-aac8-e44ac6febe79" [ 767.407339] env[62000]: _type = "Task" [ 767.407339] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.416413] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524307c4-2e1d-dee2-aac8-e44ac6febe79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.425882] env[62000]: DEBUG oslo_vmware.api [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882123, 'name': PowerOnVM_Task, 'duration_secs': 1.183785} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.426276] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 767.426728] env[62000]: INFO nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Took 9.13 seconds to spawn the instance on the hypervisor. [ 767.427101] env[62000]: DEBUG nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.427951] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-498d96af-c1e2-4599-aa5e-77650bf05d44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.441735] env[62000]: DEBUG nova.compute.manager [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Received event network-vif-plugged-c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 767.444442] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] Acquiring lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.444731] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.444909] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 767.445377] env[62000]: DEBUG nova.compute.manager [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] No waiting events found dispatching network-vif-plugged-c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 767.449314] env[62000]: WARNING nova.compute.manager [req-5f582d9b-7929-4dd8-bd31-31b16b637167 req-0e558cba-bce7-4641-b91c-5a981ae055f3 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Received unexpected event network-vif-plugged-c5103916-cf86-45f6-a1f6-286d08868edd for instance with vm_state building and task_state spawning. [ 767.549042] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Successfully updated port: c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 767.587838] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882126, 'name': ReconfigVM_Task, 'duration_secs': 0.562533} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.588166] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Reconfigured VM instance instance-00000035 to attach disk [datastore1] df525129-0ccb-4863-8a22-dd3e5a1aa2b5/df525129-0ccb-4863-8a22-dd3e5a1aa2b5.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 767.588846] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00221072-74f4-41b1-8b0d-7ead5f259a09 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.597715] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 767.597715] env[62000]: value = "task-882127" [ 767.597715] env[62000]: _type = "Task" [ 767.597715] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.606762] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882127, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.739021] env[62000]: DEBUG nova.scheduler.client.report [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 767.923043] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524307c4-2e1d-dee2-aac8-e44ac6febe79, 'name': SearchDatastore_Task, 'duration_secs': 0.023674} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.923043] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b24e1dd-672c-4429-a8cb-250f888570f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.928104] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 767.928104] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522c85fa-b739-9969-4b07-659e60074f3d" [ 767.928104] env[62000]: _type = "Task" [ 767.928104] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.938093] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522c85fa-b739-9969-4b07-659e60074f3d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.960648] env[62000]: INFO nova.compute.manager [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Took 33.10 seconds to build instance. [ 768.053119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.053119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquired lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.054181] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 768.110904] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882127, 'name': Rename_Task, 'duration_secs': 0.210374} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.113531] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 768.113531] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf76885c-465a-470e-a4c9-febf3e726747 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.122024] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 768.122024] env[62000]: value = "task-882128" [ 768.122024] env[62000]: _type = "Task" [ 768.122024] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.134185] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882128, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.245019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.510s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.245019] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.246598] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.282s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.247363] env[62000]: DEBUG nova.objects.instance [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lazy-loading 'resources' on Instance uuid 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 768.302450] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3134ea15-2faa-423a-b2d1-b242281e421b tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.323s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.441642] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522c85fa-b739-9969-4b07-659e60074f3d, 'name': SearchDatastore_Task, 'duration_secs': 0.012175} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.442331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.443172] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4c49fdd0-2485-4791-9349-a79a8663bbc0/4c49fdd0-2485-4791-9349-a79a8663bbc0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 768.443721] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3e046c1a-3e39-4f61-8ca3-68ce43a98db4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.453467] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 768.453467] env[62000]: value = "task-882129" [ 768.453467] env[62000]: _type = "Task" [ 768.453467] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.463607] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7a649414-399f-43a5-9abe-9703a56f8682 tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 96.133s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.463607] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.602108] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.641696] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882128, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.751229] env[62000]: DEBUG nova.compute.utils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.755727] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.756135] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 768.864084] env[62000]: DEBUG nova.compute.manager [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 768.864084] env[62000]: DEBUG nova.compute.manager [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing instance network info cache due to event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 768.864365] env[62000]: DEBUG oslo_concurrency.lockutils [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.864426] env[62000]: DEBUG oslo_concurrency.lockutils [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.865156] env[62000]: DEBUG nova.network.neutron [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 768.883146] env[62000]: DEBUG nova.policy [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'afdefedc02bc4b3e9b92336f57a9997b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e45b26d92ce540cea0f7cd27f52da2ec', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 768.953728] env[62000]: DEBUG nova.network.neutron [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Updating instance_info_cache with network_info: [{"id": "c5103916-cf86-45f6-a1f6-286d08868edd", "address": "fa:16:3e:eb:24:37", "network": {"id": "ea5d4361-56ca-4bdc-98b2-91aaa6601dda", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1727010769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4db75d84722b4f5faf90c531ffee3865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5103916-cf", "ovs_interfaceid": "c5103916-cf86-45f6-a1f6-286d08868edd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.975994] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882129, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.135893] env[62000]: DEBUG oslo_vmware.api [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882128, 'name': PowerOnVM_Task, 'duration_secs': 0.61122} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.135893] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 769.136905] env[62000]: INFO nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Took 8.37 seconds to spawn the instance on the hypervisor. [ 769.136905] env[62000]: DEBUG nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.137137] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc2d497f-ab40-4a8c-a6db-087db6158c4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.225015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de536b75-6038-49a4-bbe1-d7c417a33600 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.234740] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77062b5d-3bbe-4066-9aeb-cf378eebd24e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.270525] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.278427] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5cc05c-bb01-4462-820b-2793bfae3e8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.285792] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51279121-a812-483a-beef-00911aa0575a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.306021] env[62000]: DEBUG nova.compute.provider_tree [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 769.376756] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Successfully created port: d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.462347] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Releasing lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.463270] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Instance network_info: |[{"id": "c5103916-cf86-45f6-a1f6-286d08868edd", "address": "fa:16:3e:eb:24:37", "network": {"id": "ea5d4361-56ca-4bdc-98b2-91aaa6601dda", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1727010769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4db75d84722b4f5faf90c531ffee3865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5103916-cf", "ovs_interfaceid": "c5103916-cf86-45f6-a1f6-286d08868edd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 769.463270] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:24:37', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c5103916-cf86-45f6-a1f6-286d08868edd', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 769.472527] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Creating folder: Project (4db75d84722b4f5faf90c531ffee3865). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.477443] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c52f009f-eb88-48db-8a76-c1331191718b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.480505] env[62000]: DEBUG nova.compute.manager [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Received event network-changed-c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 769.480659] env[62000]: DEBUG nova.compute.manager [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Refreshing instance network info cache due to event network-changed-c5103916-cf86-45f6-a1f6-286d08868edd. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 769.480973] env[62000]: DEBUG oslo_concurrency.lockutils [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] Acquiring lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.481139] env[62000]: DEBUG oslo_concurrency.lockutils [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] Acquired lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.481378] env[62000]: DEBUG nova.network.neutron [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Refreshing network info cache for port c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 769.482819] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57685} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.483324] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4c49fdd0-2485-4791-9349-a79a8663bbc0/4c49fdd0-2485-4791-9349-a79a8663bbc0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 769.483546] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 769.484268] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0032effc-5300-4091-8018-ef4fbfd8981c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.492944] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 769.492944] env[62000]: value = "task-882131" [ 769.492944] env[62000]: _type = "Task" [ 769.492944] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.498780] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Created folder: Project (4db75d84722b4f5faf90c531ffee3865) in parent group-v201431. [ 769.499020] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Creating folder: Instances. Parent ref: group-v201522. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 769.499685] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-29d25fa0-16be-4566-90f2-738b252ac158 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.508077] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882131, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.520522] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Created folder: Instances in parent group-v201522. [ 769.520843] env[62000]: DEBUG oslo.service.loopingcall [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 769.521070] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 769.521664] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-541ce4cf-0714-4b3c-b131-d696b0601af8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.549678] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 769.549678] env[62000]: value = "task-882133" [ 769.549678] env[62000]: _type = "Task" [ 769.549678] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.561539] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882133, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.666923] env[62000]: INFO nova.compute.manager [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Took 34.46 seconds to build instance. [ 769.810403] env[62000]: DEBUG nova.scheduler.client.report [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.005248] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882131, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.098651} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.006146] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 770.006909] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cd28bdb-b342-458c-b22e-0a2628bf734a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.039316] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Reconfiguring VM instance instance-00000036 to attach disk [datastore2] 4c49fdd0-2485-4791-9349-a79a8663bbc0/4c49fdd0-2485-4791-9349-a79a8663bbc0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 770.039684] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584dbde9-2cd2-44f5-a34b-5d3d2486c04a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.068978] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882133, 'name': CreateVM_Task, 'duration_secs': 0.367214} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.069197] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 770.070185] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.070379] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.070887] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 770.072447] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b8b8fce4-39e0-425b-900d-109a996aafbb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.074704] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 770.074704] env[62000]: value = "task-882134" [ 770.074704] env[62000]: _type = "Task" [ 770.074704] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.080286] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 770.080286] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522028b3-6fa3-f27b-9272-b6b7c5ea8e12" [ 770.080286] env[62000]: _type = "Task" [ 770.080286] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.087519] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882134, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.094222] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522028b3-6fa3-f27b-9272-b6b7c5ea8e12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.127102] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.127376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.169779] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f2d4473f-6ed4-4f54-81bb-be7277f2478b tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.849s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.209055] env[62000]: DEBUG nova.network.neutron [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updated VIF entry in instance network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.209454] env[62000]: DEBUG nova.network.neutron [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.286252] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.318150] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.318678] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.319025] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.319345] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.319725] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.320103] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.320495] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.320787] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.321128] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.321560] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.321892] env[62000]: DEBUG nova.virt.hardware [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.322890] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.076s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.327108] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e59719ef-955b-402a-8695-f5f20c83d05f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.332536] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.629s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.332890] env[62000]: DEBUG nova.objects.instance [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 770.351304] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f53f1d-148f-40bd-88ef-e5211a565de6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.356237] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Successfully created port: 765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 770.370426] env[62000]: INFO nova.scheduler.client.report [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Deleted allocations for instance 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0 [ 770.498730] env[62000]: DEBUG nova.network.neutron [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Updated VIF entry in instance network info cache for port c5103916-cf86-45f6-a1f6-286d08868edd. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 770.499233] env[62000]: DEBUG nova.network.neutron [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Updating instance_info_cache with network_info: [{"id": "c5103916-cf86-45f6-a1f6-286d08868edd", "address": "fa:16:3e:eb:24:37", "network": {"id": "ea5d4361-56ca-4bdc-98b2-91aaa6601dda", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1727010769-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4db75d84722b4f5faf90c531ffee3865", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7d689fd7-f53e-4fd3-80d9-8d6b8fb7a164", "external-id": "nsx-vlan-transportzone-972", "segmentation_id": 972, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc5103916-cf", "ovs_interfaceid": "c5103916-cf86-45f6-a1f6-286d08868edd", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.587815] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882134, 'name': ReconfigVM_Task, 'duration_secs': 0.40313} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.588527] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Reconfigured VM instance instance-00000036 to attach disk [datastore2] 4c49fdd0-2485-4791-9349-a79a8663bbc0/4c49fdd0-2485-4791-9349-a79a8663bbc0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 770.589236] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-590b45b2-b813-4267-8536-0d56bd8306eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.594569] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522028b3-6fa3-f27b-9272-b6b7c5ea8e12, 'name': SearchDatastore_Task, 'duration_secs': 0.015429} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.595282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.595528] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 770.595768] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.595921] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.596146] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 770.596789] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4147a93b-52e6-4a86-8fc2-e3a4cd912a77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.600704] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 770.600704] env[62000]: value = "task-882135" [ 770.600704] env[62000]: _type = "Task" [ 770.600704] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.605601] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 770.605695] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 770.606772] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cff3ce99-7e63-4bf0-a399-b0b8dca94410 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.613201] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882135, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.616918] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 770.616918] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52215d76-4d1f-a2f3-f880-9cd7161b9fb5" [ 770.616918] env[62000]: _type = "Task" [ 770.616918] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.626468] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52215d76-4d1f-a2f3-f880-9cd7161b9fb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.633303] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 770.712521] env[62000]: DEBUG oslo_concurrency.lockutils [req-f2409d05-3e11-4ee5-a35b-976f4f93cacf req-e1d3dcf6-53c0-47ef-a166-e5468b9cca52 service nova] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.885970] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2de16c1-b59d-47bd-9551-62f67ff48677 tempest-ServerRescueTestJSONUnderV235-773521553 tempest-ServerRescueTestJSONUnderV235-773521553-project-member] Lock "64ff4a7b-ec89-48cd-8fb6-124e0726d6f0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.218s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.978765] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.978865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.979071] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 770.980045] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 770.980045] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.982144] env[62000]: INFO nova.compute.manager [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Terminating instance [ 770.984338] env[62000]: DEBUG nova.compute.manager [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.984569] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 770.985475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c084856-1739-4dcc-af58-58896ecadfb1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.994659] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 770.994935] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-52280c92-837f-4b3f-a0a1-b51d2f719fc6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.002765] env[62000]: DEBUG oslo_concurrency.lockutils [req-af5579f0-549c-4a7d-9649-ad22cf79b35e req-431d2956-c374-47c7-85f2-c17ce98fafa1 service nova] Releasing lock "refresh_cache-2b0c7354-1622-4318-ba09-6e3214eed4db" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.003182] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 771.003182] env[62000]: value = "task-882136" [ 771.003182] env[62000]: _type = "Task" [ 771.003182] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.023130] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882136, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.115782] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882135, 'name': Rename_Task, 'duration_secs': 0.343039} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.116442] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 771.116892] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fd5f61b-e2d2-44e9-b4c3-9b31cc5e7a50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.129956] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52215d76-4d1f-a2f3-f880-9cd7161b9fb5, 'name': SearchDatastore_Task, 'duration_secs': 0.021233} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.132356] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 771.132356] env[62000]: value = "task-882137" [ 771.132356] env[62000]: _type = "Task" [ 771.132356] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.132625] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c98246c-7f21-44c3-8622-6f95b74a77cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.148521] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 771.148521] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a89c7b-925a-35fc-0e16-79ac6da029f7" [ 771.148521] env[62000]: _type = "Task" [ 771.148521] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.157342] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882137, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.164090] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a89c7b-925a-35fc-0e16-79ac6da029f7, 'name': SearchDatastore_Task, 'duration_secs': 0.014578} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.164463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.164745] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2b0c7354-1622-4318-ba09-6e3214eed4db/2b0c7354-1622-4318-ba09-6e3214eed4db.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 771.166994] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.167874] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16888b9f-4e27-4004-b426-e33a8c2945e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.177441] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 771.177441] env[62000]: value = "task-882138" [ 771.177441] env[62000]: _type = "Task" [ 771.177441] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.189138] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882138, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.345327] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5aa39b19-0f9f-4d8d-bfca-a3c90f79332e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.348614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.614s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.349222] env[62000]: DEBUG nova.objects.instance [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lazy-loading 'resources' on Instance uuid af49714d-8e50-4159-96a5-cf8f70580471 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 771.516657] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882136, 'name': PowerOffVM_Task, 'duration_secs': 0.325028} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 771.517009] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 771.517135] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 771.517411] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9014453b-ea1f-458d-9712-542e32dcb82e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.594433] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 771.594853] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 771.595587] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Deleting the datastore file [datastore1] df525129-0ccb-4863-8a22-dd3e5a1aa2b5 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 771.600516] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-faffce3f-67b6-42b5-9bca-7eb18d676bca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.612077] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for the task: (returnval){ [ 771.612077] env[62000]: value = "task-882140" [ 771.612077] env[62000]: _type = "Task" [ 771.612077] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.621912] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.654313] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882137, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.698770] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882138, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.129278] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.150485] env[62000]: DEBUG oslo_vmware.api [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882137, 'name': PowerOnVM_Task, 'duration_secs': 0.953423} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.150485] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 772.150652] env[62000]: INFO nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Took 8.87 seconds to spawn the instance on the hypervisor. [ 772.150831] env[62000]: DEBUG nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.151816] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b226d552-b14a-4058-a7e2-f4969d569f58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.191682] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882138, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638015} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.191682] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2b0c7354-1622-4318-ba09-6e3214eed4db/2b0c7354-1622-4318-ba09-6e3214eed4db.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 772.191920] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 772.192217] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-88e94861-524c-4bb1-8556-0516ba44db3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.201105] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 772.201105] env[62000]: value = "task-882141" [ 772.201105] env[62000]: _type = "Task" [ 772.201105] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.216633] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882141, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.287611] env[62000]: DEBUG nova.compute.manager [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-vif-plugged-d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 772.288742] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] Acquiring lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.288986] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.002s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.289253] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.290560] env[62000]: DEBUG nova.compute.manager [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] No waiting events found dispatching network-vif-plugged-d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 772.290560] env[62000]: WARNING nova.compute.manager [req-6a225b02-4cc5-41f7-a3a0-68f29a59bb69 req-e41721ff-1a57-4657-898a-62b672c15ca3 service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received unexpected event network-vif-plugged-d210205e-eb47-430b-bbcc-5d1edc2443b9 for instance with vm_state building and task_state spawning. [ 772.309666] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f261db9-55ea-48b4-96eb-cbbc487942b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.318890] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e4f395-a777-4f3a-9288-1bae8ecad42a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.365152] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85b4be66-f2aa-4100-bccd-6a3d16882743 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.377095] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-934b130b-389a-48bf-a01a-0e31bdc2ed5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.397566] env[62000]: DEBUG nova.compute.provider_tree [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 772.625272] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882140, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.644533] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Successfully updated port: d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 772.678395] env[62000]: INFO nova.compute.manager [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Took 33.17 seconds to build instance. [ 772.715280] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882141, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.182157} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.715766] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 772.716672] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e511e1e7-f6ce-4fcc-8c46-bcff8fc1ac54 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.744884] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Reconfiguring VM instance instance-00000037 to attach disk [datastore2] 2b0c7354-1622-4318-ba09-6e3214eed4db/2b0c7354-1622-4318-ba09-6e3214eed4db.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 772.745571] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-131deb3a-68bc-4977-aeae-217e5221da60 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.768831] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 772.768831] env[62000]: value = "task-882142" [ 772.768831] env[62000]: _type = "Task" [ 772.768831] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.901082] env[62000]: DEBUG nova.scheduler.client.report [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 773.127486] env[62000]: DEBUG oslo_vmware.api [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Task: {'id': task-882140, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.058409} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.128034] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 773.128370] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 773.128679] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 773.128974] env[62000]: INFO nova.compute.manager [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Took 2.14 seconds to destroy the instance on the hypervisor. [ 773.129458] env[62000]: DEBUG oslo.service.loopingcall [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 773.132024] env[62000]: DEBUG nova.compute.manager [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.132024] env[62000]: DEBUG nova.network.neutron [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 773.184885] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7d82ece-7dbc-4382-9c12-62fe8b633ca9 tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.805s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.285472] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882142, 'name': ReconfigVM_Task, 'duration_secs': 0.300756} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.285472] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Reconfigured VM instance instance-00000037 to attach disk [datastore2] 2b0c7354-1622-4318-ba09-6e3214eed4db/2b0c7354-1622-4318-ba09-6e3214eed4db.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 773.287517] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e5ff28d-ff86-4dba-8598-c4ec8c58c3bd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.295432] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 773.295432] env[62000]: value = "task-882143" [ 773.295432] env[62000]: _type = "Task" [ 773.295432] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.306680] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882143, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.411889] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.064s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.418019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.137s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.419380] env[62000]: INFO nova.compute.claims [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.447434] env[62000]: INFO nova.scheduler.client.report [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Deleted allocations for instance af49714d-8e50-4159-96a5-cf8f70580471 [ 773.806938] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882143, 'name': Rename_Task, 'duration_secs': 0.159438} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.807290] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 773.807565] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-72f2250f-f599-4de7-892a-5323c02d6e7f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.816240] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 773.816240] env[62000]: value = "task-882144" [ 773.816240] env[62000]: _type = "Task" [ 773.816240] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.826749] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882144, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.956301] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b3e61df-7e96-49a1-8ba5-f011ecd21b57 tempest-AttachInterfacesV270Test-1769623561 tempest-AttachInterfacesV270Test-1769623561-project-member] Lock "af49714d-8e50-4159-96a5-cf8f70580471" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.384s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.252050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.252405] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.252629] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.252824] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.253062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.255620] env[62000]: INFO nova.compute.manager [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Terminating instance [ 774.260825] env[62000]: DEBUG nova.compute.manager [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.262016] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 774.262016] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c772cef3-0724-4428-adb3-b7369c9d2324 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.273854] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 774.274257] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7def0e5a-db46-4e1d-9159-7df408594d63 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.282740] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 774.282740] env[62000]: value = "task-882145" [ 774.282740] env[62000]: _type = "Task" [ 774.282740] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.295848] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-882145, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.327601] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882144, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.531430] env[62000]: DEBUG nova.compute.manager [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-changed-d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 774.531674] env[62000]: DEBUG nova.compute.manager [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Refreshing instance network info cache due to event network-changed-d210205e-eb47-430b-bbcc-5d1edc2443b9. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 774.532123] env[62000]: DEBUG oslo_concurrency.lockutils [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] Acquiring lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.532317] env[62000]: DEBUG oslo_concurrency.lockutils [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] Acquired lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.532518] env[62000]: DEBUG nova.network.neutron [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Refreshing network info cache for port d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 774.537903] env[62000]: DEBUG nova.network.neutron [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.795300] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-882145, 'name': PowerOffVM_Task, 'duration_secs': 0.333242} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.795697] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 774.795758] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 774.796037] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a1e4841f-9364-49c8-87f5-b7424802cdcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.829914] env[62000]: DEBUG oslo_vmware.api [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882144, 'name': PowerOnVM_Task, 'duration_secs': 0.723806} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.830063] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 774.830311] env[62000]: INFO nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Took 8.04 seconds to spawn the instance on the hypervisor. [ 774.830866] env[62000]: DEBUG nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 774.831542] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771acfd0-870b-42ce-a9b9-0b90f33d9c28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.835971] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc433575-4aa9-4049-97b3-46156b31de8f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.846502] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d909f1-e671-483f-84a6-61f38e9a0b4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.889253] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5629a034-a87e-4d08-87f8-5af768d96e06 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.893340] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 774.893672] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 774.893772] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Deleting the datastore file [datastore2] 210af329-4cdb-4c3f-9e82-e72a2ea79421 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 774.894083] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3736590e-3fef-4cce-a84f-86db8192c06e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.903349] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23638842-916b-4d91-8390-0b0887533b68 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.908273] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for the task: (returnval){ [ 774.908273] env[62000]: value = "task-882147" [ 774.908273] env[62000]: _type = "Task" [ 774.908273] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.921133] env[62000]: DEBUG nova.compute.provider_tree [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 774.929089] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-882147, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.042293] env[62000]: INFO nova.compute.manager [-] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Took 1.91 seconds to deallocate network for instance. [ 775.086239] env[62000]: DEBUG nova.network.neutron [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 775.151707] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.153568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.153568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.153568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.153568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.155842] env[62000]: INFO nova.compute.manager [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Terminating instance [ 775.161023] env[62000]: DEBUG nova.compute.manager [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 775.161023] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 775.161023] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8d31037b-1fb1-4172-92df-874dd6af99fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.173899] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 775.173899] env[62000]: value = "task-882148" [ 775.173899] env[62000]: _type = "Task" [ 775.173899] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.188361] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882148, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.365652] env[62000]: DEBUG nova.network.neutron [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.370471] env[62000]: INFO nova.compute.manager [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Took 32.52 seconds to build instance. [ 775.421639] env[62000]: DEBUG oslo_vmware.api [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Task: {'id': task-882147, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249361} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.421995] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 775.422269] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 775.422534] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 775.422731] env[62000]: INFO nova.compute.manager [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Took 1.16 seconds to destroy the instance on the hypervisor. [ 775.424048] env[62000]: DEBUG oslo.service.loopingcall [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 775.424048] env[62000]: DEBUG nova.compute.manager [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.424048] env[62000]: DEBUG nova.network.neutron [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 775.425768] env[62000]: DEBUG nova.scheduler.client.report [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 775.550620] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.686083] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882148, 'name': PowerOffVM_Task, 'duration_secs': 0.22334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.686453] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 775.686721] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 775.686958] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201440', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'name': 'volume-5684c204-cdf0-4896-8023-0152376f3fb3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '334fd514-c7b5-402f-935e-1d95f9b5dbc8', 'attached_at': '', 'detached_at': '', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'serial': '5684c204-cdf0-4896-8023-0152376f3fb3'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 775.687813] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff172d77-1c16-46cd-8899-52a50e8c52b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.714123] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb4061bd-9836-4fda-9993-ca683b6c52d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.726730] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87b47d7c-0a78-4c12-9a64-959d6a382994 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.749079] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b30dadf-51e0-4fe4-81d2-2044374afead {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.771227] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] The volume has not been displaced from its original location: [datastore2] volume-5684c204-cdf0-4896-8023-0152376f3fb3/volume-5684c204-cdf0-4896-8023-0152376f3fb3.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 775.779999] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Reconfiguring VM instance instance-0000002a to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 775.779999] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-afd24d8d-dcc1-4f31-86c4-71e54155d851 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.803452] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 775.803452] env[62000]: value = "task-882149" [ 775.803452] env[62000]: _type = "Task" [ 775.803452] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.813282] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882149, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.871481] env[62000]: DEBUG oslo_concurrency.lockutils [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] Releasing lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.871789] env[62000]: DEBUG nova.compute.manager [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Received event network-vif-deleted-0c52b803-0943-434c-a133-1f05adb65a75 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 775.872009] env[62000]: INFO nova.compute.manager [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Neutron deleted interface 0c52b803-0943-434c-a133-1f05adb65a75; detaching it from the instance and deleting it from the info cache [ 775.872222] env[62000]: DEBUG nova.network.neutron [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.875774] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2aa04a68-7da3-41b6-bf88-913d29be0005 tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.287s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.932022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.516s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.932349] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 775.936021] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.357s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.936021] env[62000]: DEBUG nova.objects.instance [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 776.315369] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882149, 'name': ReconfigVM_Task, 'duration_secs': 0.353739} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.315755] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Reconfigured VM instance instance-0000002a to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 776.321653] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b43ab3f3-39bd-4e5d-9fb6-6369a34eaaae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.337015] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Successfully updated port: 765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 776.347420] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 776.347420] env[62000]: value = "task-882150" [ 776.347420] env[62000]: _type = "Task" [ 776.347420] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.358151] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882150, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.376901] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da065f3d-c785-4304-ad6b-06c0b476e901 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.388900] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16331b34-f4cf-454b-b590-00986ca2b24d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.425104] env[62000]: DEBUG nova.compute.manager [req-992c1fd3-4b0e-41d5-b5fc-dcf21f80f16e req-42413305-ee65-4285-961d-4d02c7ec144c service nova] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Detach interface failed, port_id=0c52b803-0943-434c-a133-1f05adb65a75, reason: Instance df525129-0ccb-4863-8a22-dd3e5a1aa2b5 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 776.440748] env[62000]: DEBUG nova.compute.utils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 776.445962] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 776.449314] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 776.784022] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 776.785006] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919580d2-8012-4d00-80c8-711484a961fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.802488] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 776.802488] env[62000]: ERROR oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk due to incomplete transfer. [ 776.802488] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1de61af5-07be-4e82-bd7f-7b0eb3764a7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.810782] env[62000]: DEBUG oslo_vmware.rw_handles [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/523698f6-f961-79a8-8df7-8d1b9b8291cb/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 776.811020] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploaded image 150ffec2-eb6a-4b9b-baf4-6005539ac9fa to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 776.813367] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 776.813588] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-996bf5e7-1746-448b-bf78-793c183c2547 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.822988] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 776.822988] env[62000]: value = "task-882151" [ 776.822988] env[62000]: _type = "Task" [ 776.822988] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.836027] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882151, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.841167] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.841227] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.842764] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.856780] env[62000]: DEBUG nova.policy [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '00cf68b6bf82491394734fc898f2c088', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9f3c844c28b745c0a824f223bba6fc39', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 776.865011] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882150, 'name': ReconfigVM_Task, 'duration_secs': 0.191191} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.865812] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201440', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'name': 'volume-5684c204-cdf0-4896-8023-0152376f3fb3', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '334fd514-c7b5-402f-935e-1d95f9b5dbc8', 'attached_at': '', 'detached_at': '', 'volume_id': '5684c204-cdf0-4896-8023-0152376f3fb3', 'serial': '5684c204-cdf0-4896-8023-0152376f3fb3'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 776.865812] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 776.866587] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b00ab66-1329-42e0-a38e-b43b813ddaa1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.876558] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 776.876821] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4b5e579d-8b3a-480d-80b7-02480ec10636 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.955239] env[62000]: DEBUG oslo_concurrency.lockutils [None req-019a4573-60e3-4783-94a7-1fe3b073303f tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.956018] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 776.962612] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.378s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.964723] env[62000]: INFO nova.compute.claims [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.991260] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 776.991599] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 776.991842] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Deleting the datastore file [datastore2] 334fd514-c7b5-402f-935e-1d95f9b5dbc8 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 776.992963] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-257cd310-daae-45c1-bd0a-75301a13a3cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.006460] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for the task: (returnval){ [ 777.006460] env[62000]: value = "task-882153" [ 777.006460] env[62000]: _type = "Task" [ 777.006460] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.016888] env[62000]: DEBUG nova.compute.manager [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-vif-plugged-765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.017203] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Acquiring lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.017757] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.017757] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.017903] env[62000]: DEBUG nova.compute.manager [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] No waiting events found dispatching network-vif-plugged-765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 777.018913] env[62000]: WARNING nova.compute.manager [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received unexpected event network-vif-plugged-765ab3cf-243a-415e-944f-c00efe23e3e0 for instance with vm_state building and task_state spawning. [ 777.018913] env[62000]: DEBUG nova.compute.manager [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-changed-765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 777.018913] env[62000]: DEBUG nova.compute.manager [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Refreshing instance network info cache due to event network-changed-765ab3cf-243a-415e-944f-c00efe23e3e0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 777.018913] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Acquiring lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.030766] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882153, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.334374] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882151, 'name': Destroy_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.438038] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.517336] env[62000]: DEBUG oslo_vmware.api [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Task: {'id': task-882153, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106666} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.518153] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 777.518357] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 777.518536] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 777.518717] env[62000]: INFO nova.compute.manager [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Took 2.36 seconds to destroy the instance on the hypervisor. [ 777.518956] env[62000]: DEBUG oslo.service.loopingcall [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 777.519171] env[62000]: DEBUG nova.compute.manager [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 777.519267] env[62000]: DEBUG nova.network.neutron [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.540465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "2b0c7354-1622-4318-ba09-6e3214eed4db" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.540465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.540465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.540465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.540465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.542207] env[62000]: INFO nova.compute.manager [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Terminating instance [ 777.545526] env[62000]: DEBUG nova.compute.manager [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 777.545728] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 777.546588] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcd1e81a-660d-4018-a794-fe1add742ec6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.562555] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 777.562954] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0ce3851d-684e-4bec-a81d-0e4f65671fce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.572127] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 777.572127] env[62000]: value = "task-882154" [ 777.572127] env[62000]: _type = "Task" [ 777.572127] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.584242] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882154, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.650441] env[62000]: DEBUG nova.network.neutron [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.837025] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882151, 'name': Destroy_Task, 'duration_secs': 0.588453} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 777.839193] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroyed the VM [ 777.839524] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 777.840063] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-008cc268-fc0a-4514-b3f7-63ebe56a6f1b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.850935] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 777.850935] env[62000]: value = "task-882155" [ 777.850935] env[62000]: _type = "Task" [ 777.850935] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 777.863230] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882155, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.982837] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 778.021908] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 778.022221] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 778.022387] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 778.022573] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 778.022776] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 778.022863] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 778.025055] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 778.025055] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 778.025055] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 778.028355] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 778.028355] env[62000]: DEBUG nova.virt.hardware [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 778.028355] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0d8203-d36c-4de0-a3a5-1b9715ecd769 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.047614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a211e4b-d607-44cd-868a-30b95c5f9d17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.076871] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Successfully created port: d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 778.085535] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882154, 'name': PowerOffVM_Task, 'duration_secs': 0.341963} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.085718] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 778.086275] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 778.086275] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-56a043ab-fc6a-4dec-9e87-76ac1be55740 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.154813] env[62000]: INFO nova.compute.manager [-] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Took 2.73 seconds to deallocate network for instance. [ 778.163594] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 778.163961] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 778.164302] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Deleting the datastore file [datastore2] 2b0c7354-1622-4318-ba09-6e3214eed4db {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 778.164691] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ca5c1d24-872c-42f8-96b7-5fab05b9b38d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.178558] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for the task: (returnval){ [ 778.178558] env[62000]: value = "task-882157" [ 778.178558] env[62000]: _type = "Task" [ 778.178558] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.191869] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882157, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.368628] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882155, 'name': RemoveSnapshot_Task} progress is 17%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.453283] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e589e62b-70bb-4777-8333-cb02a664db74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.464305] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b7b84ea-39f7-476f-ae61-840e63661f07 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.511559] env[62000]: DEBUG nova.network.neutron [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [{"id": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "address": "fa:16:3e:93:54:a3", "network": {"id": "c8a86e03-aea5-4b12-8fab-748dc138e07d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1345451530", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd210205e-eb", "ovs_interfaceid": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "765ab3cf-243a-415e-944f-c00efe23e3e0", "address": "fa:16:3e:36:f9:3e", "network": {"id": "c873698e-2fee-4f1c-bab5-4c17a496d6c9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1660274167", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap765ab3cf-24", "ovs_interfaceid": "765ab3cf-243a-415e-944f-c00efe23e3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.513589] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28fae51a-754c-4ad0-a1f3-dbce49aa8160 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.523205] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d67550-9808-4bfc-bf31-63871f53756a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.540741] env[62000]: DEBUG nova.compute.provider_tree [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.656152] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.656152] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.656724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.656724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.656903] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.660106] env[62000]: INFO nova.compute.manager [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Terminating instance [ 778.662029] env[62000]: DEBUG nova.compute.manager [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 778.662262] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 778.665568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.665568] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d091e20-8f16-4dc1-9ed8-ccf1687b3c09 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.672849] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 778.673159] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f2e4e44-5cf9-4e58-81c1-317eba186c4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.683605] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 778.683605] env[62000]: value = "task-882158" [ 778.683605] env[62000]: _type = "Task" [ 778.683605] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 778.694843] env[62000]: DEBUG oslo_vmware.api [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Task: {'id': task-882157, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278524} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.695585] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 778.695783] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 778.695984] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 778.696497] env[62000]: INFO nova.compute.manager [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Took 1.15 seconds to destroy the instance on the hypervisor. [ 778.697160] env[62000]: DEBUG oslo.service.loopingcall [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 778.697160] env[62000]: DEBUG nova.compute.manager [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.697160] env[62000]: DEBUG nova.network.neutron [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 778.704515] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882158, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.866973] env[62000]: DEBUG oslo_vmware.api [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882155, 'name': RemoveSnapshot_Task, 'duration_secs': 0.990969} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.867379] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 778.867624] env[62000]: INFO nova.compute.manager [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 17.81 seconds to snapshot the instance on the hypervisor. [ 779.020076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.020490] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance network_info: |[{"id": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "address": "fa:16:3e:93:54:a3", "network": {"id": "c8a86e03-aea5-4b12-8fab-748dc138e07d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1345451530", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd210205e-eb", "ovs_interfaceid": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "765ab3cf-243a-415e-944f-c00efe23e3e0", "address": "fa:16:3e:36:f9:3e", "network": {"id": "c873698e-2fee-4f1c-bab5-4c17a496d6c9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1660274167", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap765ab3cf-24", "ovs_interfaceid": "765ab3cf-243a-415e-944f-c00efe23e3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 779.020837] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Acquired lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.021037] env[62000]: DEBUG nova.network.neutron [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Refreshing network info cache for port 765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 779.022308] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:93:54:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b49c5024-2ced-42ca-90cc-6066766d43e6', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd210205e-eb47-430b-bbcc-5d1edc2443b9', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:36:f9:3e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c80243e-93a7-4a95-bc8d-e9534bacd66e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '765ab3cf-243a-415e-944f-c00efe23e3e0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 779.032656] env[62000]: DEBUG oslo.service.loopingcall [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.033234] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 779.033464] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d36af329-e26e-49f1-96a9-f351d18e6d4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.054687] env[62000]: DEBUG nova.scheduler.client.report [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.066075] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 779.066075] env[62000]: value = "task-882159" [ 779.066075] env[62000]: _type = "Task" [ 779.066075] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.076138] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882159, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.201499] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882158, 'name': PowerOffVM_Task, 'duration_secs': 0.230206} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.201891] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 779.202132] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 779.202518] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b7dfaf96-c641-446b-938a-21d67eaee9d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.279439] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 779.280924] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 779.280924] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Deleting the datastore file [datastore2] 4c49fdd0-2485-4791-9349-a79a8663bbc0 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 779.280924] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dd340f23-ba57-4e86-9a06-cc611b7d7a81 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.291117] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for the task: (returnval){ [ 779.291117] env[62000]: value = "task-882161" [ 779.291117] env[62000]: _type = "Task" [ 779.291117] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.301149] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.436616] env[62000]: DEBUG nova.network.neutron [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.491565] env[62000]: DEBUG nova.compute.manager [None req-165dfbab-5c7c-496b-b17f-afe4b947fa6a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Found 1 images (rotation: 2) {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 779.546507] env[62000]: DEBUG nova.compute.manager [req-e9eebbb2-05ab-45fc-bc50-4c661545df61 req-71fa1131-2983-498f-979d-cef29344aa8e service nova] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Received event network-vif-deleted-892944d2-5b6b-4a93-b975-f9e1f8104b1e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.546507] env[62000]: DEBUG nova.compute.manager [req-e9eebbb2-05ab-45fc-bc50-4c661545df61 req-71fa1131-2983-498f-979d-cef29344aa8e service nova] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Received event network-vif-deleted-d89856ae-9afc-460b-bc77-5a3635aa59f7 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 779.560546] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.598s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.561291] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.564687] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.219s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.564687] env[62000]: DEBUG nova.objects.instance [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lazy-loading 'resources' on Instance uuid 4e617bb5-84e4-4505-99e3-61289826f511 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.580920] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882159, 'name': CreateVM_Task, 'duration_secs': 0.458806} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.583837] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 779.583837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.583837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.583837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 779.587620] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5de2db1a-0533-45ce-a8d6-1cc866f324db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.590985] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 779.590985] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524e8c52-31fd-d279-f0fc-7a1e45a19edb" [ 779.590985] env[62000]: _type = "Task" [ 779.590985] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.606268] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524e8c52-31fd-d279-f0fc-7a1e45a19edb, 'name': SearchDatastore_Task, 'duration_secs': 0.010169} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.606268] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.606268] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 779.609955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.609955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.609955] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 779.609955] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-120a2995-4df8-47ce-990c-ae86b29b4ffe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.619717] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 779.619717] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 779.619717] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00de560e-f319-4d12-bba1-f34c00178b80 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.626133] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 779.626133] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528751f9-fce4-3b42-beec-cf6d3dfecab9" [ 779.626133] env[62000]: _type = "Task" [ 779.626133] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 779.637299] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528751f9-fce4-3b42-beec-cf6d3dfecab9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 779.795427] env[62000]: DEBUG nova.network.neutron [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.805301] env[62000]: DEBUG oslo_vmware.api [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Task: {'id': task-882161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212065} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 779.805301] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 779.805301] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 779.805301] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 779.805301] env[62000]: INFO nova.compute.manager [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Took 1.14 seconds to destroy the instance on the hypervisor. [ 779.811855] env[62000]: DEBUG oslo.service.loopingcall [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 779.811855] env[62000]: DEBUG nova.compute.manager [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 779.811855] env[62000]: DEBUG nova.network.neutron [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 779.941110] env[62000]: INFO nova.compute.manager [-] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Took 2.42 seconds to deallocate network for instance. [ 780.077115] env[62000]: DEBUG nova.compute.utils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 780.079809] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 780.079993] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 780.139802] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528751f9-fce4-3b42-beec-cf6d3dfecab9, 'name': SearchDatastore_Task, 'duration_secs': 0.026438} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.147177] env[62000]: DEBUG nova.policy [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4770851cc3b4ad8aaa6866a18ee1359', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41e10f7d4f964f2795dc629721802880', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.148562] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb2d7f8a-3772-4213-b859-b0aadf624647 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.159837] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 780.159837] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c0cefd-c0cb-9506-2f41-d4ceb10d4fec" [ 780.159837] env[62000]: _type = "Task" [ 780.159837] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.167623] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c0cefd-c0cb-9506-2f41-d4ceb10d4fec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.287517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.287517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.295641] env[62000]: DEBUG nova.network.neutron [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updated VIF entry in instance network info cache for port 765ab3cf-243a-415e-944f-c00efe23e3e0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 780.296322] env[62000]: DEBUG nova.network.neutron [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [{"id": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "address": "fa:16:3e:93:54:a3", "network": {"id": "c8a86e03-aea5-4b12-8fab-748dc138e07d", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1345451530", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b49c5024-2ced-42ca-90cc-6066766d43e6", "external-id": "nsx-vlan-transportzone-239", "segmentation_id": 239, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd210205e-eb", "ovs_interfaceid": "d210205e-eb47-430b-bbcc-5d1edc2443b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "765ab3cf-243a-415e-944f-c00efe23e3e0", "address": "fa:16:3e:36:f9:3e", "network": {"id": "c873698e-2fee-4f1c-bab5-4c17a496d6c9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1660274167", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap765ab3cf-24", "ovs_interfaceid": "765ab3cf-243a-415e-944f-c00efe23e3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.298162] env[62000]: INFO nova.compute.manager [-] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Took 1.60 seconds to deallocate network for instance. [ 780.523993] env[62000]: INFO nova.compute.manager [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Took 0.58 seconds to detach 1 volumes for instance. [ 780.527914] env[62000]: DEBUG nova.compute.manager [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Deleting volume: 5684c204-cdf0-4896-8023-0152376f3fb3 {{(pid=62000) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 780.531981] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3f9fa9-6038-416a-a0d2-20a5b215d257 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.541581] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4b18bd-b105-49d4-8a48-dcc9cd99d49c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.589918] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.596410] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839f2397-49ba-4bdb-b9b5-e64d7d46f309 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.605318] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37aa140e-4dc9-4ed9-b4b5-b6afb9fe9fe0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.621035] env[62000]: DEBUG nova.compute.provider_tree [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.652144] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Successfully updated port: d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 780.664064] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Successfully created port: 00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.673151] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c0cefd-c0cb-9506-2f41-d4ceb10d4fec, 'name': SearchDatastore_Task, 'duration_secs': 0.013561} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 780.673701] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.673701] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 79a0d6f0-7dca-4c5d-8656-91bf4d957e78/79a0d6f0-7dca-4c5d-8656-91bf4d957e78.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 780.674297] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52682cca-771c-4c2a-8bd1-e625f2acc525 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.683597] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 780.683597] env[62000]: value = "task-882163" [ 780.683597] env[62000]: _type = "Task" [ 780.683597] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 780.692340] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882163, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 780.790433] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 780.799465] env[62000]: DEBUG oslo_concurrency.lockutils [req-22fac49b-dd3c-446a-b55b-279578961904 req-cc63d1ce-f57e-4ce3-b62b-41f23a9cafac service nova] Releasing lock "refresh_cache-79a0d6f0-7dca-4c5d-8656-91bf4d957e78" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.813286] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.994975] env[62000]: DEBUG nova.network.neutron [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.103187] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.130403] env[62000]: DEBUG nova.scheduler.client.report [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.154213] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.154376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquired lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.154531] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 781.195372] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882163, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481932} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.195973] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 79a0d6f0-7dca-4c5d-8656-91bf4d957e78/79a0d6f0-7dca-4c5d-8656-91bf4d957e78.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 781.196085] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 781.196414] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-883dac58-0005-495a-b5c1-da29dd45a11e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.204939] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 781.204939] env[62000]: value = "task-882164" [ 781.204939] env[62000]: _type = "Task" [ 781.204939] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.215816] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882164, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.314069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.500807] env[62000]: INFO nova.compute.manager [-] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Took 1.69 seconds to deallocate network for instance. [ 781.544101] env[62000]: DEBUG nova.compute.manager [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 781.545043] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d559ed-160c-4c3f-9789-f33582ab1754 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.579292] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Received event network-vif-deleted-c5103916-cf86-45f6-a1f6-286d08868edd {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.579481] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Received event network-vif-plugged-d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.579681] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Acquiring lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.579939] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.580187] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.580366] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] No waiting events found dispatching network-vif-plugged-d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 781.580540] env[62000]: WARNING nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Received unexpected event network-vif-plugged-d35218f5-dff3-4425-972f-7baa9578e319 for instance with vm_state building and task_state spawning. [ 781.580710] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Received event network-vif-deleted-b6da36aa-e361-46b2-99a3-7c942b5ae516 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.580915] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Received event network-changed-d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 781.581096] env[62000]: DEBUG nova.compute.manager [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Refreshing instance network info cache due to event network-changed-d35218f5-dff3-4425-972f-7baa9578e319. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 781.581273] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Acquiring lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.604777] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.632289] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.632580] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.632751] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.632961] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.633143] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.633300] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.633512] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.633687] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.633857] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.634035] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.634408] env[62000]: DEBUG nova.virt.hardware [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.635089] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3047f298-568a-4ca7-b48e-72280f403a16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.638169] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.074s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.640217] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 26.442s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.640445] env[62000]: DEBUG nova.objects.instance [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'resources' on Instance uuid 596af5ab-1791-42ce-93d2-3e4f0a47dfa2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 781.647591] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eccc3de7-41f8-4136-a7b5-e1b8f7b2ddc1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.665972] env[62000]: INFO nova.scheduler.client.report [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Deleted allocations for instance 4e617bb5-84e4-4505-99e3-61289826f511 [ 781.697162] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.715652] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882164, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068594} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.715946] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 781.720513] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f61d80a-06ea-4203-9027-12a30c569d81 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.748050] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Reconfiguring VM instance instance-00000038 to attach disk [datastore2] 79a0d6f0-7dca-4c5d-8656-91bf4d957e78/79a0d6f0-7dca-4c5d-8656-91bf4d957e78.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 781.748702] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b9ed9c4-471c-448a-be7f-28bfdb2c1f5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.786790] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 781.786790] env[62000]: value = "task-882165" [ 781.786790] env[62000]: _type = "Task" [ 781.786790] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.796038] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882165, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.928401] env[62000]: DEBUG nova.network.neutron [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Updating instance_info_cache with network_info: [{"id": "d35218f5-dff3-4425-972f-7baa9578e319", "address": "fa:16:3e:24:80:ea", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd35218f5-df", "ovs_interfaceid": "d35218f5-dff3-4425-972f-7baa9578e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.007782] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.057595] env[62000]: INFO nova.compute.manager [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] instance snapshotting [ 782.058256] env[62000]: DEBUG nova.objects.instance [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 782.174413] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c7f2fd40-2a09-4593-9d7b-847ab508a84e tempest-SecurityGroupsTestJSON-1610690152 tempest-SecurityGroupsTestJSON-1610690152-project-member] Lock "4e617bb5-84e4-4505-99e3-61289826f511" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 31.977s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.299072] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882165, 'name': ReconfigVM_Task, 'duration_secs': 0.284731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.299072] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Reconfigured VM instance instance-00000038 to attach disk [datastore2] 79a0d6f0-7dca-4c5d-8656-91bf4d957e78/79a0d6f0-7dca-4c5d-8656-91bf4d957e78.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 782.299259] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2cb73d08-8359-4b57-8d9e-18eac7f996af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.307443] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 782.307443] env[62000]: value = "task-882166" [ 782.307443] env[62000]: _type = "Task" [ 782.307443] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.320709] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882166, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.435360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Releasing lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.435676] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Instance network_info: |[{"id": "d35218f5-dff3-4425-972f-7baa9578e319", "address": "fa:16:3e:24:80:ea", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd35218f5-df", "ovs_interfaceid": "d35218f5-dff3-4425-972f-7baa9578e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 782.436299] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Acquired lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.436552] env[62000]: DEBUG nova.network.neutron [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Refreshing network info cache for port d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 782.437682] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:80:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7c09cc47-a7d0-4816-bee4-69cc9f2e04b0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd35218f5-dff3-4425-972f-7baa9578e319', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 782.444983] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Creating folder: Project (9f3c844c28b745c0a824f223bba6fc39). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.448202] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aaa9c7b7-1b45-4610-ac5a-39c871e64dec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.471089] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Created folder: Project (9f3c844c28b745c0a824f223bba6fc39) in parent group-v201431. [ 782.471336] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Creating folder: Instances. Parent ref: group-v201526. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 782.471594] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8daef78a-9afe-401f-96bf-6885abd3d1c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.486872] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Created folder: Instances in parent group-v201526. [ 782.486872] env[62000]: DEBUG oslo.service.loopingcall [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.489042] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612775cc-fc1b-4b1b-b403-8c2b66182477 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.492891] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 782.493432] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b953fe2f-264e-430f-a849-d1ceb136cd38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.517625] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf918b98-1ac8-41f4-b9bb-c77a356cb2e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.521891] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 782.521891] env[62000]: value = "task-882169" [ 782.521891] env[62000]: _type = "Task" [ 782.521891] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.553113] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b7b01d-a079-4b42-8a22-a7d9560c96b1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.560016] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882169, 'name': CreateVM_Task} progress is 15%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.566195] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0033cdbb-cf92-4cf3-89c4-fa5841d68712 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.571326] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fec8b361-2148-4e47-874c-2d6a5d72921e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.598719] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-380f71c0-a0a0-4c94-89ce-84e3d59ee25f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.601550] env[62000]: DEBUG nova.compute.provider_tree [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.640138] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Successfully updated port: 00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 782.679724] env[62000]: DEBUG nova.compute.manager [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Received event network-vif-plugged-00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 782.680155] env[62000]: DEBUG oslo_concurrency.lockutils [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] Acquiring lock "ae343199-1006-473a-a47f-7983835bd60a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.680562] env[62000]: DEBUG oslo_concurrency.lockutils [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] Lock "ae343199-1006-473a-a47f-7983835bd60a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.680652] env[62000]: DEBUG oslo_concurrency.lockutils [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] Lock "ae343199-1006-473a-a47f-7983835bd60a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.681231] env[62000]: DEBUG nova.compute.manager [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] No waiting events found dispatching network-vif-plugged-00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 782.681485] env[62000]: WARNING nova.compute.manager [req-393b6f4a-89c9-4ddd-8d35-a2be2c849e8d req-63cfa5ef-7c49-4b69-bfca-eca7d8987bcf service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Received unexpected event network-vif-plugged-00896a83-2b1c-4c70-86a9-243a6ccbc36f for instance with vm_state building and task_state spawning. [ 782.821320] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882166, 'name': Rename_Task, 'duration_secs': 0.16614} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.821681] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 782.822589] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f2d6df66-5ced-4878-95ad-2d487f187ded {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.830218] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 782.830218] env[62000]: value = "task-882170" [ 782.830218] env[62000]: _type = "Task" [ 782.830218] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 782.838490] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882170, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.036360] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882169, 'name': CreateVM_Task, 'duration_secs': 0.492479} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.036493] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 783.037822] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.038203] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.038624] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 783.039200] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9da5e9db-83f0-4bb7-a316-b6ac76610dee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.046940] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 783.046940] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e1394a-380b-3bd9-d586-f635c7614ba2" [ 783.046940] env[62000]: _type = "Task" [ 783.046940] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.061610] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e1394a-380b-3bd9-d586-f635c7614ba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.104348] env[62000]: DEBUG nova.scheduler.client.report [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.113452] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 783.113862] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4b2fa649-df9c-4f73-8547-6f7c264cde6b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.127379] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 783.127379] env[62000]: value = "task-882171" [ 783.127379] env[62000]: _type = "Task" [ 783.127379] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.138681] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882171, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.142441] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.145018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.145018] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 783.270994] env[62000]: DEBUG nova.network.neutron [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Updated VIF entry in instance network info cache for port d35218f5-dff3-4425-972f-7baa9578e319. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 783.271421] env[62000]: DEBUG nova.network.neutron [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Updating instance_info_cache with network_info: [{"id": "d35218f5-dff3-4425-972f-7baa9578e319", "address": "fa:16:3e:24:80:ea", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.68", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd35218f5-df", "ovs_interfaceid": "d35218f5-dff3-4425-972f-7baa9578e319", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.348023] env[62000]: DEBUG oslo_vmware.api [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882170, 'name': PowerOnVM_Task, 'duration_secs': 0.474007} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.348023] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 783.348023] env[62000]: INFO nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Took 13.06 seconds to spawn the instance on the hypervisor. [ 783.348023] env[62000]: DEBUG nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 783.348023] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870410f2-a5a7-4a15-b20f-d3b2bd1696ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.560091] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e1394a-380b-3bd9-d586-f635c7614ba2, 'name': SearchDatastore_Task, 'duration_secs': 0.010474} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 783.560466] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.560718] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 783.560965] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.561139] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.561327] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 783.561607] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-536bfe88-a127-43f6-9cfd-cbb3d602c5d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.571806] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 783.572082] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 783.572836] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd9bd0af-78b5-4942-8f90-67f6ed2edd20 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.579017] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 783.579017] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8e0a-bd61-8a52-906d-888ef6e1c5fb" [ 783.579017] env[62000]: _type = "Task" [ 783.579017] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 783.587343] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8e0a-bd61-8a52-906d-888ef6e1c5fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.609486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.969s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.614325] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.766s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.618265] env[62000]: INFO nova.compute.claims [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.639886] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882171, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 783.648603] env[62000]: INFO nova.scheduler.client.report [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocations for instance 596af5ab-1791-42ce-93d2-3e4f0a47dfa2 [ 783.713548] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.778882] env[62000]: DEBUG oslo_concurrency.lockutils [req-4907c4fe-d4ac-4e65-935d-6bb6d5943ccc req-64e2f7a7-47af-4549-b99e-2617433dc54f service nova] Releasing lock "refresh_cache-296d5fea-e9d2-49e3-bc29-8e4bf3782535" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.868398] env[62000]: INFO nova.compute.manager [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Took 40.43 seconds to build instance. [ 784.005620] env[62000]: DEBUG nova.network.neutron [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Updating instance_info_cache with network_info: [{"id": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "address": "fa:16:3e:5a:df:25", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00896a83-2b", "ovs_interfaceid": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.090835] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8e0a-bd61-8a52-906d-888ef6e1c5fb, 'name': SearchDatastore_Task, 'duration_secs': 0.025007} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.091672] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13c0a578-de12-4c3b-a657-adf8a8085b10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.098023] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 784.098023] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52478e5d-1f1f-5b1e-9e05-b46d7ec57f67" [ 784.098023] env[62000]: _type = "Task" [ 784.098023] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.107959] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52478e5d-1f1f-5b1e-9e05-b46d7ec57f67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.140115] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882171, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.164780] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f3dabec5-5c21-4d10-a435-4486413b0e47 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "596af5ab-1791-42ce-93d2-3e4f0a47dfa2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 33.312s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.370107] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0b6a98c6-c111-4709-81d6-217f41b06e01 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.852s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.510341] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.510701] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance network_info: |[{"id": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "address": "fa:16:3e:5a:df:25", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00896a83-2b", "ovs_interfaceid": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 784.511168] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:df:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00896a83-2b1c-4c70-86a9-243a6ccbc36f', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 784.521855] env[62000]: DEBUG oslo.service.loopingcall [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.522183] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 784.522420] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fa25a298-6021-42e1-a01d-496b6252e7d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.556227] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 784.556227] env[62000]: value = "task-882172" [ 784.556227] env[62000]: _type = "Task" [ 784.556227] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.563985] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882172, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.610979] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52478e5d-1f1f-5b1e-9e05-b46d7ec57f67, 'name': SearchDatastore_Task, 'duration_secs': 0.019545} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.611714] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.612026] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 296d5fea-e9d2-49e3-bc29-8e4bf3782535/296d5fea-e9d2-49e3-bc29-8e4bf3782535.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 784.612313] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-934db4fe-2017-4af3-8505-ff5c2afd272a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.619091] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 784.619091] env[62000]: value = "task-882173" [ 784.619091] env[62000]: _type = "Task" [ 784.619091] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.627750] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 784.639878] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882171, 'name': CreateSnapshot_Task, 'duration_secs': 1.222675} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 784.640990] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 784.647023] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a938853-0d1d-4e67-a2ed-87258dc6ae90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.770718] env[62000]: DEBUG nova.compute.manager [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Received event network-changed-00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 784.771777] env[62000]: DEBUG nova.compute.manager [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Refreshing instance network info cache due to event network-changed-00896a83-2b1c-4c70-86a9-243a6ccbc36f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 784.771777] env[62000]: DEBUG oslo_concurrency.lockutils [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] Acquiring lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.771777] env[62000]: DEBUG oslo_concurrency.lockutils [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] Acquired lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.771777] env[62000]: DEBUG nova.network.neutron [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Refreshing network info cache for port 00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 784.854226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.854226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.854226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.854226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.854226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.858318] env[62000]: INFO nova.compute.manager [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Terminating instance [ 784.860531] env[62000]: DEBUG nova.compute.manager [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 784.860746] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 784.861704] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3706febd-fc83-45b5-b459-3c6641fb59a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.873046] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 784.873429] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1d8d6797-3428-4791-b7fd-326ea684b220 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.879830] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 784.879830] env[62000]: value = "task-882174" [ 784.879830] env[62000]: _type = "Task" [ 784.879830] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 784.890918] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882174, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.043333] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d868503a-abe2-4ba3-858a-e4513260dc26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.051856] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2caf42-9eed-4797-9725-8f9f919b8fc2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.088625] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0420ee64-3944-4cc9-9708-5c394c5a3cad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.093905] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882172, 'name': CreateVM_Task, 'duration_secs': 0.349348} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.094608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 785.095415] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.095594] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.096143] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 785.098975] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d6b4d88-16af-4644-a0ca-0b2159dbe426 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.102207] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-623fb925-85a1-46fd-a7e1-819bb858460f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.119645] env[62000]: DEBUG nova.compute.provider_tree [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.122266] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 785.122266] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f9082-dab0-93a3-aa5e-707061bae08e" [ 785.122266] env[62000]: _type = "Task" [ 785.122266] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.135554] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882173, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.140999] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f9082-dab0-93a3-aa5e-707061bae08e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.165409] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 785.165753] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-430d0681-7b7f-4ac9-81b2-31193e66de15 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.175499] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 785.175499] env[62000]: value = "task-882175" [ 785.175499] env[62000]: _type = "Task" [ 785.175499] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.184286] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882175, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.390205] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882174, 'name': PowerOffVM_Task, 'duration_secs': 0.389994} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.390497] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 785.390664] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 785.391209] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f491aa68-2797-4291-9653-5982e6ce9133 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.515553] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 785.515800] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 785.516090] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleting the datastore file [datastore2] 79a0d6f0-7dca-4c5d-8656-91bf4d957e78 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 785.516288] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-92bff6a2-6bd5-4683-a4c3-87ca3114d94f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.523139] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for the task: (returnval){ [ 785.523139] env[62000]: value = "task-882177" [ 785.523139] env[62000]: _type = "Task" [ 785.523139] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.535381] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882177, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.628511] env[62000]: DEBUG nova.scheduler.client.report [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.648635] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f9082-dab0-93a3-aa5e-707061bae08e, 'name': SearchDatastore_Task, 'duration_secs': 0.027418} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.648926] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542022} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.649247] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.649501] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 785.649764] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.649944] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.650159] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 785.650439] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 296d5fea-e9d2-49e3-bc29-8e4bf3782535/296d5fea-e9d2-49e3-bc29-8e4bf3782535.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 785.650803] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 785.651206] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6b0e3d55-8dfe-435d-bcc9-ff363876d96c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.653312] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0e7c6dbb-ad03-4eb3-9020-f35aaf3affa8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.662884] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 785.662884] env[62000]: value = "task-882178" [ 785.662884] env[62000]: _type = "Task" [ 785.662884] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.669167] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 785.669434] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 785.672206] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-411fb65c-b95f-400b-b32d-d17576f82ad5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.678772] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882178, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.684211] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 785.684211] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ae87c2-2192-9a84-857a-2ef0096b6191" [ 785.684211] env[62000]: _type = "Task" [ 785.684211] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.691207] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882175, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.702208] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ae87c2-2192-9a84-857a-2ef0096b6191, 'name': SearchDatastore_Task, 'duration_secs': 0.013332} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 785.702208] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22708455-51cc-4ab7-b8d6-8c64fc42de5b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.704888] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 785.704888] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238bee4-ede0-d8c2-7022-6c70f8162954" [ 785.704888] env[62000]: _type = "Task" [ 785.704888] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 785.713049] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238bee4-ede0-d8c2-7022-6c70f8162954, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 785.804532] env[62000]: DEBUG nova.network.neutron [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Updated VIF entry in instance network info cache for port 00896a83-2b1c-4c70-86a9-243a6ccbc36f. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 785.805044] env[62000]: DEBUG nova.network.neutron [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Updating instance_info_cache with network_info: [{"id": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "address": "fa:16:3e:5a:df:25", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap00896a83-2b", "ovs_interfaceid": "00896a83-2b1c-4c70-86a9-243a6ccbc36f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.039125] env[62000]: DEBUG oslo_vmware.api [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Task: {'id': task-882177, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.451767} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.039307] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 786.041245] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 786.041245] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 786.041245] env[62000]: INFO nova.compute.manager [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Took 1.18 seconds to destroy the instance on the hypervisor. [ 786.041245] env[62000]: DEBUG oslo.service.loopingcall [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 786.041245] env[62000]: DEBUG nova.compute.manager [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 786.041245] env[62000]: DEBUG nova.network.neutron [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 786.134990] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.135565] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.139955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.364s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.140244] env[62000]: DEBUG nova.objects.instance [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lazy-loading 'resources' on Instance uuid a9783a2a-a7d3-4fbe-9590-d530ac24fa82 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 786.180154] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882178, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064137} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.183772] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 786.185122] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-658099f1-f747-4e32-99f3-62358de84370 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.193150] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882175, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.215216] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 296d5fea-e9d2-49e3-bc29-8e4bf3782535/296d5fea-e9d2-49e3-bc29-8e4bf3782535.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 786.215978] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5724bf5d-2904-4f97-ab24-c736188ca0a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.241732] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238bee4-ede0-d8c2-7022-6c70f8162954, 'name': SearchDatastore_Task, 'duration_secs': 0.017238} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 786.243179] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.243387] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 786.244032] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 786.244032] env[62000]: value = "task-882179" [ 786.244032] env[62000]: _type = "Task" [ 786.244032] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.244217] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5377586d-4cf5-490e-ac29-bd8cb2f150ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.254385] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882179, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.256250] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 786.256250] env[62000]: value = "task-882180" [ 786.256250] env[62000]: _type = "Task" [ 786.256250] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 786.264204] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882180, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.307639] env[62000]: DEBUG oslo_concurrency.lockutils [req-36e9907e-fa39-4abe-86a9-2e6a79f9b1ef req-f9d9c591-e3de-41ba-8186-0afb535ac039 service nova] Releasing lock "refresh_cache-ae343199-1006-473a-a47f-7983835bd60a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.644526] env[62000]: DEBUG nova.compute.utils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.652661] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.653760] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 786.696759] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882175, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.758068] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882179, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.770873] env[62000]: DEBUG nova.policy [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffd8327dd8c4844b776982e99de5e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14aec090f2a4abc8078df1fe4428bbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.780129] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882180, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 786.886429] env[62000]: DEBUG nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-vif-deleted-d210205e-eb47-430b-bbcc-5d1edc2443b9 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 786.887161] env[62000]: INFO nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Neutron deleted interface d210205e-eb47-430b-bbcc-5d1edc2443b9; detaching it from the instance and deleting it from the info cache [ 786.888388] env[62000]: DEBUG nova.network.neutron [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [{"id": "765ab3cf-243a-415e-944f-c00efe23e3e0", "address": "fa:16:3e:36:f9:3e", "network": {"id": "c873698e-2fee-4f1c-bab5-4c17a496d6c9", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1660274167", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.205", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "e45b26d92ce540cea0f7cd27f52da2ec", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c80243e-93a7-4a95-bc8d-e9534bacd66e", "external-id": "nsx-vlan-transportzone-306", "segmentation_id": 306, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap765ab3cf-24", "ovs_interfaceid": "765ab3cf-243a-415e-944f-c00efe23e3e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.154151] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.198192] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882175, 'name': CloneVM_Task, 'duration_secs': 1.755051} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.198192] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created linked-clone VM from snapshot [ 787.198192] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54668d78-25a9-4cec-8b12-56b6d23bc49a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.204270] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploading image 68dcb5e3-03ff-4018-8972-717abd7984a7 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 787.232510] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 787.232510] env[62000]: value = "vm-201531" [ 787.232510] env[62000]: _type = "VirtualMachine" [ 787.232510] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 787.232785] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bda863d4-b988-4b6c-93ef-4985fe889a37 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.236351] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c3ab6d-0f3c-4819-a0b9-95aa85c5bbe4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.242922] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40119d24-501e-4d0f-b9ce-37be40471626 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.248080] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease: (returnval){ [ 787.248080] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52868c58-8d92-efd1-ceb1-332e943481d6" [ 787.248080] env[62000]: _type = "HttpNfcLease" [ 787.248080] env[62000]: } obtained for exporting VM: (result){ [ 787.248080] env[62000]: value = "vm-201531" [ 787.248080] env[62000]: _type = "VirtualMachine" [ 787.248080] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 787.248316] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the lease: (returnval){ [ 787.248316] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52868c58-8d92-efd1-ceb1-332e943481d6" [ 787.248316] env[62000]: _type = "HttpNfcLease" [ 787.248316] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 787.281857] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Successfully created port: 7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.286684] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c6fe00-0764-4a66-8eb3-749984b869de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.294140] env[62000]: DEBUG nova.network.neutron [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.295722] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882179, 'name': ReconfigVM_Task, 'duration_secs': 0.543848} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.295929] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 787.295929] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52868c58-8d92-efd1-ceb1-332e943481d6" [ 787.295929] env[62000]: _type = "HttpNfcLease" [ 787.295929] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 787.300020] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 296d5fea-e9d2-49e3-bc29-8e4bf3782535/296d5fea-e9d2-49e3-bc29-8e4bf3782535.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 787.300020] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 787.300020] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52868c58-8d92-efd1-ceb1-332e943481d6" [ 787.300020] env[62000]: _type = "HttpNfcLease" [ 787.300020] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 787.300020] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-af6833c3-d467-464f-8e60-58490ee04557 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.300020] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b8314d0-dd99-485c-aa21-fca3c0d38068 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.307295] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882180, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.672681} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.309619] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a9d003e-8b40-4592-ab10-0f3ce7a956fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.313595] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 787.313824] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 787.318335] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5545340b-65c4-4252-a202-509b433beee6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.320302] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 787.320302] env[62000]: value = "task-882185" [ 787.320302] env[62000]: _type = "Task" [ 787.320302] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.320550] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 787.320722] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 787.337321] env[62000]: DEBUG nova.compute.provider_tree [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 787.396867] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 787.396867] env[62000]: value = "task-882186" [ 787.396867] env[62000]: _type = "Task" [ 787.396867] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.398924] env[62000]: DEBUG nova.scheduler.client.report [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.405483] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a24e6346-7449-4f07-a1f0-2586eaf4baa4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.407672] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882185, 'name': Rename_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.418309] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882186, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074446} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.419478] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 787.420345] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2439630-d68a-49d0-b06c-ae90ae32fee7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.425970] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a144c0c-5e37-4f99-9197-536337ded3d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.460059] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Reconfiguring VM instance instance-0000003a to attach disk [datastore1] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 787.470642] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad95d005-ecce-4971-80e7-7cfef3a592ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.485745] env[62000]: DEBUG nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Detach interface failed, port_id=d210205e-eb47-430b-bbcc-5d1edc2443b9, reason: Instance 79a0d6f0-7dca-4c5d-8656-91bf4d957e78 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 787.485980] env[62000]: DEBUG nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Received event network-vif-deleted-765ab3cf-243a-415e-944f-c00efe23e3e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 787.486173] env[62000]: INFO nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Neutron deleted interface 765ab3cf-243a-415e-944f-c00efe23e3e0; detaching it from the instance and deleting it from the info cache [ 787.486350] env[62000]: DEBUG nova.network.neutron [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.487751] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-68b7aa18-1e46-44eb-88ed-63dde3a804dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.495218] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 787.495218] env[62000]: value = "task-882187" [ 787.495218] env[62000]: _type = "Task" [ 787.495218] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.506339] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882187, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.798460] env[62000]: INFO nova.compute.manager [-] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Took 1.76 seconds to deallocate network for instance. [ 787.836894] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882185, 'name': Rename_Task, 'duration_secs': 0.143476} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 787.838208] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 787.839512] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-80dbf026-013d-4a5b-9aca-7f36bf7f307b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.845808] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 787.845808] env[62000]: value = "task-882188" [ 787.845808] env[62000]: _type = "Task" [ 787.845808] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 787.854789] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882188, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 787.910233] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.769s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.911903] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 28.773s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.940513] env[62000]: INFO nova.scheduler.client.report [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Deleted allocations for instance a9783a2a-a7d3-4fbe-9590-d530ac24fa82 [ 787.992538] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1655d915-84c3-4aa2-8caa-89989b4dfd66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.006223] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e917e7-1d77-4176-97db-2700c65a9f6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.020875] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.043860] env[62000]: DEBUG nova.compute.manager [req-77e6b0a5-1ed6-4577-96d6-15f9661294fb req-466d9e92-3f02-4919-aad7-1554d8677fff service nova] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Detach interface failed, port_id=765ab3cf-243a-415e-944f-c00efe23e3e0, reason: Instance 79a0d6f0-7dca-4c5d-8656-91bf4d957e78 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 788.163914] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.200130] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.200351] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.204136] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.204393] env[62000]: DEBUG nova.virt.hardware [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.205164] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328a6a83-31e8-49b4-bef7-58adda5751c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.217714] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6975985-493f-4954-8966-7c8a61956256 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.308567] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.356828] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882188, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.421612] env[62000]: INFO nova.compute.claims [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 788.451574] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a84cb73a-a630-41c2-ab55-6906f2261916 tempest-ServerShowV257Test-1284019296 tempest-ServerShowV257Test-1284019296-project-member] Lock "a9783a2a-a7d3-4fbe-9590-d530ac24fa82" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.562s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.512745] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882187, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.793694] env[62000]: DEBUG nova.compute.manager [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Received event network-vif-plugged-7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 788.794008] env[62000]: DEBUG oslo_concurrency.lockutils [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] Acquiring lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.794387] env[62000]: DEBUG oslo_concurrency.lockutils [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 788.794700] env[62000]: DEBUG oslo_concurrency.lockutils [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.795039] env[62000]: DEBUG nova.compute.manager [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] No waiting events found dispatching network-vif-plugged-7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 788.795284] env[62000]: WARNING nova.compute.manager [req-24d87323-b01a-45d8-89af-68f3b6f925cd req-539ff285-3706-4eed-8894-57151fab5401 service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Received unexpected event network-vif-plugged-7e374e87-d2b5-49ef-9516-991307309c6d for instance with vm_state building and task_state spawning. [ 788.863086] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882188, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 788.934176] env[62000]: INFO nova.compute.resource_tracker [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating resource usage from migration 5e0eeb7d-c7f1-4c3a-b730-71573fb528fc [ 789.013547] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882187, 'name': ReconfigVM_Task, 'duration_secs': 1.488332} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.014288] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Reconfigured VM instance instance-0000003a to attach disk [datastore1] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 789.015077] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d036e1bb-7c81-4fa9-b4fe-32392942e80c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.022820] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 789.022820] env[62000]: value = "task-882189" [ 789.022820] env[62000]: _type = "Task" [ 789.022820] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.033245] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882189, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.073631] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Successfully updated port: 7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 789.346935] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d1f24c-c54d-4efc-809b-081b3d1f5e7f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.363179] env[62000]: DEBUG oslo_vmware.api [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882188, 'name': PowerOnVM_Task, 'duration_secs': 1.238037} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.365734] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 789.366114] env[62000]: INFO nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Took 11.39 seconds to spawn the instance on the hypervisor. [ 789.366314] env[62000]: DEBUG nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 789.367651] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bec8eda-7412-4a53-bb09-5658aaaebeb0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.371234] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caa0aa4f-2b08-448c-b6c1-6dca4a75d106 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.409761] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd308d43-dcd6-4e85-a440-9ced051004eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.418852] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a881088-6985-43a0-95e4-1f7722c969e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.432586] env[62000]: DEBUG nova.compute.provider_tree [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.535423] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882189, 'name': Rename_Task, 'duration_secs': 0.167124} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 789.535719] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 789.535970] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b232849-ea2d-432a-91a3-b769b3d6d470 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.543942] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 789.543942] env[62000]: value = "task-882190" [ 789.543942] env[62000]: _type = "Task" [ 789.543942] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 789.554895] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882190, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 789.576938] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.577141] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.577300] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 789.894921] env[62000]: INFO nova.compute.manager [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Took 37.64 seconds to build instance. [ 789.938764] env[62000]: DEBUG nova.scheduler.client.report [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 790.010483] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.010483] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.055964] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882190, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.125430] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 790.290638] env[62000]: DEBUG nova.network.neutron [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Updating instance_info_cache with network_info: [{"id": "7e374e87-d2b5-49ef-9516-991307309c6d", "address": "fa:16:3e:f2:d2:23", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e374e87-d2", "ovs_interfaceid": "7e374e87-d2b5-49ef-9516-991307309c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.396999] env[62000]: DEBUG oslo_concurrency.lockutils [None req-94f17a1b-b21f-4f82-8011-c39fa7c66ee1 tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.153s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.444663] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.533s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 790.444894] env[62000]: INFO nova.compute.manager [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Migrating [ 790.445153] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.445310] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.446559] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.206s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.448112] env[62000]: INFO nova.compute.claims [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 790.454022] env[62000]: INFO nova.compute.rpcapi [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 790.454022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.512734] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 790.557648] env[62000]: DEBUG oslo_vmware.api [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882190, 'name': PowerOnVM_Task, 'duration_secs': 0.692642} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 790.558657] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 790.558876] env[62000]: INFO nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Took 8.95 seconds to spawn the instance on the hypervisor. [ 790.559345] env[62000]: DEBUG nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 790.560544] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ec45274-2bf1-4a39-a29a-d5f9d7e2dbc7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.793268] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 790.793450] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Instance network_info: |[{"id": "7e374e87-d2b5-49ef-9516-991307309c6d", "address": "fa:16:3e:f2:d2:23", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e374e87-d2", "ovs_interfaceid": "7e374e87-d2b5-49ef-9516-991307309c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 790.794340] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:d2:23', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e374e87-d2b5-49ef-9516-991307309c6d', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 790.802215] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating folder: Project (d14aec090f2a4abc8078df1fe4428bbb). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.802571] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-00ef2f9b-a538-4c31-a071-bf41cd23b534 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.814541] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created folder: Project (d14aec090f2a4abc8078df1fe4428bbb) in parent group-v201431. [ 790.814874] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating folder: Instances. Parent ref: group-v201535. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 790.815009] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1ac2c794-6383-4c22-8dda-cb5d4804836b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.824753] env[62000]: DEBUG nova.compute.manager [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Received event network-changed-7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 790.824990] env[62000]: DEBUG nova.compute.manager [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Refreshing instance network info cache due to event network-changed-7e374e87-d2b5-49ef-9516-991307309c6d. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 790.825264] env[62000]: DEBUG oslo_concurrency.lockutils [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] Acquiring lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.825419] env[62000]: DEBUG oslo_concurrency.lockutils [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] Acquired lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.825611] env[62000]: DEBUG nova.network.neutron [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Refreshing network info cache for port 7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 790.828269] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created folder: Instances in parent group-v201535. [ 790.828516] env[62000]: DEBUG oslo.service.loopingcall [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.828931] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 790.829173] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3f4f307c-7eea-4fa1-979b-461825a8c042 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.852370] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 790.852370] env[62000]: value = "task-882194" [ 790.852370] env[62000]: _type = "Task" [ 790.852370] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 790.861382] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882194, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 790.972018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 790.972018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 790.972018] env[62000]: DEBUG nova.network.neutron [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 791.048939] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.082093] env[62000]: INFO nova.compute.manager [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Took 38.53 seconds to build instance. [ 791.365624] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882194, 'name': CreateVM_Task, 'duration_secs': 0.376845} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.365624] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 791.366709] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.366786] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.367368] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 791.368302] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e21427f-440d-4e5f-a382-a811abfd877f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.372536] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 791.372536] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b5f460-2ee3-e0b8-c96a-f2599c016791" [ 791.372536] env[62000]: _type = "Task" [ 791.372536] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.381619] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b5f460-2ee3-e0b8-c96a-f2599c016791, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 791.583674] env[62000]: DEBUG oslo_concurrency.lockutils [None req-31efd6d3-0ca2-44cf-b73a-f3f7db96cd9c tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.049s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.795773] env[62000]: DEBUG nova.network.neutron [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Updated VIF entry in instance network info cache for port 7e374e87-d2b5-49ef-9516-991307309c6d. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 791.797903] env[62000]: DEBUG nova.network.neutron [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Updating instance_info_cache with network_info: [{"id": "7e374e87-d2b5-49ef-9516-991307309c6d", "address": "fa:16:3e:f2:d2:23", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e374e87-d2", "ovs_interfaceid": "7e374e87-d2b5-49ef-9516-991307309c6d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.851337] env[62000]: DEBUG nova.network.neutron [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.887470] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b5f460-2ee3-e0b8-c96a-f2599c016791, 'name': SearchDatastore_Task, 'duration_secs': 0.043954} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.887863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.888152] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 791.888425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.888618] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.888788] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 791.889112] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4b649b7-c529-4d35-9cd4-202a345b492f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.902019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d926cd2-27bc-40ab-92e9-a91ac9d5e4a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.904565] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 791.904797] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 791.908472] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-045999fd-0116-444e-8db7-3c41398013fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.916611] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b938a2-1c2f-43f7-8f26-3534a579bc19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.920095] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 791.920095] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526f87b1-f5ff-a055-3481-81bc9a9cb676" [ 791.920095] env[62000]: _type = "Task" [ 791.920095] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.950721] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a2d8ec-d31c-45b5-b053-db4ce6a901ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.957342] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526f87b1-f5ff-a055-3481-81bc9a9cb676, 'name': SearchDatastore_Task, 'duration_secs': 0.012781} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.958632] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7fc77b3b-b92f-4623-8c42-7e4f2a55e5f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.964963] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f452ac72-2511-46b5-927a-73c093cab9cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.970494] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 791.970494] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e53df3-16c2-ef31-020b-53d41f9c02dc" [ 791.970494] env[62000]: _type = "Task" [ 791.970494] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 791.983347] env[62000]: DEBUG nova.compute.provider_tree [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 791.990289] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e53df3-16c2-ef31-020b-53d41f9c02dc, 'name': SearchDatastore_Task, 'duration_secs': 0.013907} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 791.990598] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.990864] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ae1e8ebc-ddb8-4e95-847e-b9684e9161d4/ae1e8ebc-ddb8-4e95-847e-b9684e9161d4.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 791.991817] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d8762470-3ef3-4379-915c-0555064dc196 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.000615] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 792.000615] env[62000]: value = "task-882195" [ 792.000615] env[62000]: _type = "Task" [ 792.000615] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.010401] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882195, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.043150] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.043293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.043631] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.043954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.043954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.052373] env[62000]: INFO nova.compute.manager [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Terminating instance [ 792.056308] env[62000]: DEBUG nova.compute.manager [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.056629] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 792.057650] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-279353f9-7e3a-4fd4-8438-ccc469ff8325 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.066735] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 792.067196] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f89e3da8-a220-45ce-a584-3f5f6f77953c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.078293] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 792.078293] env[62000]: value = "task-882197" [ 792.078293] env[62000]: _type = "Task" [ 792.078293] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.086007] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882197, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.306052] env[62000]: DEBUG oslo_concurrency.lockutils [req-3ebfe062-72be-4c30-be99-8c3c2f267e5f req-cace4ce9-58d1-4310-a900-09497843d9bd service nova] Releasing lock "refresh_cache-ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.354764] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.488561] env[62000]: DEBUG nova.scheduler.client.report [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 792.518299] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882195, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.586852] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882197, 'name': PowerOffVM_Task, 'duration_secs': 0.269374} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 792.587178] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 792.587397] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 792.587666] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-431d39f2-8b78-43af-b20e-412c0c363b72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.669190] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 792.669558] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 792.669885] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Deleting the datastore file [datastore2] 296d5fea-e9d2-49e3-bc29-8e4bf3782535 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 792.670872] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db046bd3-148b-43f1-9898-6c12b196c7ed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.679714] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for the task: (returnval){ [ 792.679714] env[62000]: value = "task-882199" [ 792.679714] env[62000]: _type = "Task" [ 792.679714] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 792.690967] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882199, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 792.748813] env[62000]: INFO nova.compute.manager [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Rebuilding instance [ 792.795998] env[62000]: DEBUG nova.compute.manager [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 792.797131] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79904509-5216-4608-9803-0b2262c98516 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.999463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.549s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 792.999463] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.000786] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 31.682s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.020203] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882195, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.5635} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.021292] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ae1e8ebc-ddb8-4e95-847e-b9684e9161d4/ae1e8ebc-ddb8-4e95-847e-b9684e9161d4.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 793.022034] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 793.022722] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7841597a-7e58-4010-9e23-e7a3866c1912 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.034479] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 793.034479] env[62000]: value = "task-882200" [ 793.034479] env[62000]: _type = "Task" [ 793.034479] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.042898] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882200, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.156505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.156802] env[62000]: DEBUG oslo_concurrency.lockutils [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.157159] env[62000]: DEBUG nova.objects.instance [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'flavor' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.191874] env[62000]: DEBUG oslo_vmware.api [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Task: {'id': task-882199, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.256903} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.192429] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 793.192865] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 793.195350] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 793.195350] env[62000]: INFO nova.compute.manager [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Took 1.14 seconds to destroy the instance on the hypervisor. [ 793.195350] env[62000]: DEBUG oslo.service.loopingcall [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.195350] env[62000]: DEBUG nova.compute.manager [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.195350] env[62000]: DEBUG nova.network.neutron [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 793.309898] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 793.310637] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8e6a945c-235a-4c42-9c64-7c586e91f8bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.319460] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 793.319460] env[62000]: value = "task-882201" [ 793.319460] env[62000]: _type = "Task" [ 793.319460] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.329288] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882201, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.502971] env[62000]: DEBUG nova.compute.utils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 793.504882] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 793.544800] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882200, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068036} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.545338] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 793.546449] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8a9459f-4598-40db-b4c1-da4540312b4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.569724] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] ae1e8ebc-ddb8-4e95-847e-b9684e9161d4/ae1e8ebc-ddb8-4e95-847e-b9684e9161d4.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 793.570820] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-139ab694-cf68-46dd-affd-f7f1e2fc73b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.591804] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 793.591804] env[62000]: value = "task-882202" [ 793.591804] env[62000]: _type = "Task" [ 793.591804] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 793.599960] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882202, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 793.661310] env[62000]: DEBUG nova.objects.instance [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'pci_requests' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.832025] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882201, 'name': PowerOffVM_Task, 'duration_secs': 0.210896} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 793.832025] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 793.832025] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 793.832645] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59fb1d0a-03d7-44e5-8054-be04d4184145 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.841414] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 793.841414] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-999877b8-19a8-4917-a4f4-24e0a3a8fbf2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.879232] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5a7116c-1a8d-41ae-b2b6-87ea0485c303 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.897302] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 793.960178] env[62000]: DEBUG nova.network.neutron [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.006708] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 794.018803] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Applying migration context for instance f3a6527c-777e-4a07-9482-598de15d4eb3 as it has an incoming, in-progress migration 5e0eeb7d-c7f1-4c3a-b730-71573fb528fc. Migration status is migrating {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 794.019374] env[62000]: INFO nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating resource usage from migration 5e0eeb7d-c7f1-4c3a-b730-71573fb528fc [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance eb27703f-b657-423a-90a9-a7c024a2e473 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 49f6c198-13b7-4c07-81d8-c010e7b0598e is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 334fd514-c7b5-402f-935e-1d95f9b5dbc8 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 210af329-4cdb-4c3f-9e82-e72a2ea79421 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 71192360-6ee6-4876-bf37-da987a09cbb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 30d82c70-1401-4a1a-a88b-f798f8fbf96a is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance ea8ae421-1d7f-4814-bc0f-90a3316ad028 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance df525129-0ccb-4863-8a22-dd3e5a1aa2b5 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 4c49fdd0-2485-4791-9349-a79a8663bbc0 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 2b0c7354-1622-4318-ba09-6e3214eed4db is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 79a0d6f0-7dca-4c5d-8656-91bf4d957e78 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 296d5fea-e9d2-49e3-bc29-8e4bf3782535 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042018] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance ae343199-1006-473a-a47f-7983835bd60a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042524] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance ae1e8ebc-ddb8-4e95-847e-b9684e9161d4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042524] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Migration 5e0eeb7d-c7f1-4c3a-b730-71573fb528fc is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 794.042524] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f3a6527c-777e-4a07-9482-598de15d4eb3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.042524] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 153b86d9-dc2e-463d-87fd-155ec23e2abb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 794.092495] env[62000]: DEBUG nova.compute.manager [req-18b91bbc-e8d1-4cfe-9d92-9245b44a9777 req-1fea587a-4c56-42ad-955c-2b731b7b786d service nova] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Received event network-vif-deleted-d35218f5-dff3-4425-972f-7baa9578e319 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 794.103307] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882202, 'name': ReconfigVM_Task, 'duration_secs': 0.291669} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.103591] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Reconfigured VM instance instance-0000003b to attach disk [datastore2] ae1e8ebc-ddb8-4e95-847e-b9684e9161d4/ae1e8ebc-ddb8-4e95-847e-b9684e9161d4.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 794.104347] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3381289-599a-4426-9957-9db7e6c4df34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.111635] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 794.111635] env[62000]: value = "task-882205" [ 794.111635] env[62000]: _type = "Task" [ 794.111635] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.124031] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882205, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.163810] env[62000]: DEBUG nova.objects.base [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Object Instance<71192360-6ee6-4876-bf37-da987a09cbb2> lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 794.164111] env[62000]: DEBUG nova.network.neutron [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 794.256573] env[62000]: DEBUG oslo_concurrency.lockutils [None req-591bc4cc-5f86-41de-a723-52ec58dbe910 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.100s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 794.403212] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 794.403638] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dc761f94-1f68-44a7-b2ef-c227dd34c7a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.410980] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.411344] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.414575] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 794.414575] env[62000]: value = "task-882206" [ 794.414575] env[62000]: _type = "Task" [ 794.414575] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.424796] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882206, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.462446] env[62000]: INFO nova.compute.manager [-] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Took 1.27 seconds to deallocate network for instance. [ 794.545678] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 96107e36-bf3f-4ef0-8d8b-5c9601f4f514 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.622244] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882205, 'name': Rename_Task, 'duration_secs': 0.139577} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.622594] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 794.622887] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c4f6fee-eb9e-4039-b429-bd602220bb7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.631374] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 794.631374] env[62000]: value = "task-882207" [ 794.631374] env[62000]: _type = "Task" [ 794.631374] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.639399] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882207, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.675126] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 794.675126] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 794.675126] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore1] ae343199-1006-473a-a47f-7983835bd60a {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 794.675370] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0a58193f-7510-4c10-a96a-78082c09681f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.683346] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 794.683346] env[62000]: value = "task-882208" [ 794.683346] env[62000]: _type = "Task" [ 794.683346] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 794.691709] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882208, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 794.916390] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 794.930239] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882206, 'name': PowerOffVM_Task, 'duration_secs': 0.20822} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 794.930239] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 794.930463] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 794.971032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.015962] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.040452] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.040775] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.041024] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.041292] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.041526] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.041765] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.042091] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.042301] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.042595] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.042773] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.042990] env[62000]: DEBUG nova.virt.hardware [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.043944] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41bca9d3-a328-4a73-b5eb-5bcd9eebd315 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.048648] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 85f5f90a-9591-4393-9bef-2ad7e6a1a82f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.054997] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc60517-bb5c-42db-8d6c-4e8d946c25f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.071111] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 795.076302] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Creating folder: Project (b08c7f33c9354fb682edcf2004e36512). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.076891] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4904aad4-1ecc-495e-9e6c-a27313068a60 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.090268] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Created folder: Project (b08c7f33c9354fb682edcf2004e36512) in parent group-v201431. [ 795.090483] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Creating folder: Instances. Parent ref: group-v201538. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 795.090724] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc7df3a1-0fde-4fda-a66f-301398d18069 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.099970] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Created folder: Instances in parent group-v201538. [ 795.100269] env[62000]: DEBUG oslo.service.loopingcall [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 795.100467] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 795.100691] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e3daad88-8f54-494c-96e8-a320652c403e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.119050] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 795.119050] env[62000]: value = "task-882211" [ 795.119050] env[62000]: _type = "Task" [ 795.119050] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.127371] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882211, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.139610] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882207, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.193514] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882208, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.230439} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.193787] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 795.193975] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 795.194211] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 795.437217] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.437922] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.437922] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.438364] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.438364] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.438639] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.438987] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.439301] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.439544] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.439879] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.440564] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.446943] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.447753] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31925f68-2483-4c75-964d-5483795ec7ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.464629] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 795.464629] env[62000]: value = "task-882212" [ 795.464629] env[62000]: _type = "Task" [ 795.464629] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.473706] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882212, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.554922] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance e2e3270d-086b-4441-a3d6-49b05a60b51f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.629788] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882211, 'name': CreateVM_Task, 'duration_secs': 0.27629} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.629788] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 795.630098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.630336] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.630704] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 795.631014] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a38107d9-fd2b-47eb-8084-ceed1f170f30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.639072] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 795.639072] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8b67a-13e2-c28a-c9e1-9de1cdec5d4d" [ 795.639072] env[62000]: _type = "Task" [ 795.639072] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 795.642615] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882207, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.650367] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8b67a-13e2-c28a-c9e1-9de1cdec5d4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 795.975342] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882212, 'name': ReconfigVM_Task, 'duration_secs': 0.211603} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 795.975770] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 796.059032] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 66edb906-f936-4f57-833d-224f36af109e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 796.059382] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 796.059426] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 796.149531] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882207, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.158225] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8b67a-13e2-c28a-c9e1-9de1cdec5d4d, 'name': SearchDatastore_Task, 'duration_secs': 0.011703} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.158438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.158696] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 796.158955] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.159123] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.159337] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 796.160033] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdd7ffb9-c48a-4cc4-8fc5-d88da7822eb1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.171046] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 796.171299] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 796.172747] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-393be328-d9f1-4c65-9415-dcdd79b43fed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.178447] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 796.178447] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522b7b06-5cad-4194-97ab-34a6b4e63533" [ 796.178447] env[62000]: _type = "Task" [ 796.178447] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.190363] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522b7b06-5cad-4194-97ab-34a6b4e63533, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.231099] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.231788] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.232156] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.232466] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.233131] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.233497] env[62000]: DEBUG nova.virt.hardware [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.236497] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02f603a8-6e11-480c-95e1-fac0b8de9e93 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.248529] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f2c5a84-3694-40cf-9274-63e344d06b38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.256389] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.256389] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.256566] env[62000]: DEBUG nova.objects.instance [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'flavor' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.266885] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5a:df:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '00896a83-2b1c-4c70-86a9-243a6ccbc36f', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 796.274810] env[62000]: DEBUG oslo.service.loopingcall [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 796.283290] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 796.283290] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5d8687de-b541-4c90-9ca7-2db6c28631be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.307714] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 796.307714] env[62000]: value = "task-882214" [ 796.307714] env[62000]: _type = "Task" [ 796.307714] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.318230] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882214, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.390938] env[62000]: DEBUG nova.objects.instance [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'pci_requests' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 796.425019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-078cc509-cdb1-4685-aa6c-665e9bc18a2c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.434062] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3294f98d-5c8d-49df-8e5e-00dc913a2625 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.468622] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d6d847-6c4d-4edb-b7d4-bf18f1b2ac1a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.478223] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c1c9e4-5c15-4aad-95f3-37f937ef1b47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.484615] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:54:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='622b3a6c-bfbe-431a-b5c4-116aa70b0e48',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-862518541',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 796.484853] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 796.485711] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 796.485711] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 796.485711] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 796.485711] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 796.485895] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 796.485929] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 796.486091] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 796.486266] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 796.486438] env[62000]: DEBUG nova.virt.hardware [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 796.491999] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfiguring VM instance instance-00000032 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 796.492936] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7d9718b-e7b0-4f5e-b17d-b1ec120d745a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.521419] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 796.523966] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 796.523966] env[62000]: value = "task-882215" [ 796.523966] env[62000]: _type = "Task" [ 796.523966] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.534032] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882215, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.546190] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 796.547470] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b14698a-afb6-4f12-8f2e-739d19eed8f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.558800] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 796.558800] env[62000]: ERROR oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk due to incomplete transfer. [ 796.560122] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-2a7a0227-7f1c-4a28-8b3e-cdc9e770d647 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.570853] env[62000]: DEBUG oslo_vmware.rw_handles [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52c30681-0979-9d76-0cc2-6f95126c959f/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 796.571097] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploaded image 68dcb5e3-03ff-4018-8972-717abd7984a7 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 796.573920] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 796.574665] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f9071ed9-daef-4377-85c0-16a7e4ce332d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.585790] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 796.585790] env[62000]: value = "task-882216" [ 796.585790] env[62000]: _type = "Task" [ 796.585790] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.598997] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882216, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.646757] env[62000]: DEBUG oslo_vmware.api [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882207, 'name': PowerOnVM_Task, 'duration_secs': 1.519078} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.647119] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 796.647346] env[62000]: INFO nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Took 8.48 seconds to spawn the instance on the hypervisor. [ 796.647542] env[62000]: DEBUG nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 796.648545] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93152e3-1157-4fd0-b804-bdc641ee6164 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.693808] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522b7b06-5cad-4194-97ab-34a6b4e63533, 'name': SearchDatastore_Task, 'duration_secs': 0.013184} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.695144] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51dfd2a9-dd7d-4e07-a465-5e8f1f348ce5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.703787] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 796.703787] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52accbca-328b-bf55-c3ca-d44d0c267b4c" [ 796.703787] env[62000]: _type = "Task" [ 796.703787] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.716483] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52accbca-328b-bf55-c3ca-d44d0c267b4c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.820878] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882214, 'name': CreateVM_Task, 'duration_secs': 0.405222} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 796.821071] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 796.821913] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 796.822262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.822626] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 796.822901] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-72cb6185-abc0-4679-b04a-66dade89a9a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.831122] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 796.831122] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52730ede-b4f3-328b-b4cb-71dad2a9fa06" [ 796.831122] env[62000]: _type = "Task" [ 796.831122] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 796.843263] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52730ede-b4f3-328b-b4cb-71dad2a9fa06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 796.894332] env[62000]: DEBUG nova.objects.base [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Object Instance<71192360-6ee6-4876-bf37-da987a09cbb2> lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 796.894630] env[62000]: DEBUG nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 796.933131] env[62000]: DEBUG nova.policy [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 797.030255] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.040904] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882215, 'name': ReconfigVM_Task, 'duration_secs': 0.219209} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.041595] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfigured VM instance instance-00000032 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 797.042381] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbebe3ac-4738-4a0b-9565-cf207a2ea8e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.065604] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 797.066569] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8ae49190-bf80-43b7-987a-0bb73dae1aec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.088471] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 797.088471] env[62000]: value = "task-882217" [ 797.088471] env[62000]: _type = "Task" [ 797.088471] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.100551] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882216, 'name': Destroy_Task, 'duration_secs': 0.378899} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.104136] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroyed the VM [ 797.104258] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 797.104532] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882217, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.104809] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e07acafb-aa1a-4881-9188-dcef8ec80f66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.112559] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 797.112559] env[62000]: value = "task-882218" [ 797.112559] env[62000]: _type = "Task" [ 797.112559] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.122507] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882218, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.172538] env[62000]: INFO nova.compute.manager [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Took 41.34 seconds to build instance. [ 797.218482] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52accbca-328b-bf55-c3ca-d44d0c267b4c, 'name': SearchDatastore_Task, 'duration_secs': 0.012339} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.219084] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.219595] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 797.220132] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe84c3db-1d81-4f01-bf93-253eba5f80b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.223384] env[62000]: DEBUG nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Successfully created port: 656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.234338] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 797.234338] env[62000]: value = "task-882219" [ 797.234338] env[62000]: _type = "Task" [ 797.234338] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.242236] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882219, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.345157] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52730ede-b4f3-328b-b4cb-71dad2a9fa06, 'name': SearchDatastore_Task, 'duration_secs': 0.01805} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.345605] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.345917] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 797.346255] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 797.346505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 797.346710] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 797.347083] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-868ae300-df91-4985-a0ad-aa43d90dea51 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.357035] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 797.357282] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 797.358155] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c1f417bc-0b46-4c23-9737-b428bbdb6f13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.364583] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 797.364583] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d8469c-50d3-a14f-6d39-6a37ebe237c2" [ 797.364583] env[62000]: _type = "Task" [ 797.364583] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.373818] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d8469c-50d3-a14f-6d39-6a37ebe237c2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.535697] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 797.536018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.535s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.536350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 34.995s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.536853] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.538898] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 31.383s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.539152] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.540914] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.374s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.542656] env[62000]: INFO nova.compute.claims [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.545353] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 797.545532] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Cleaning up deleted instances {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 797.575905] env[62000]: INFO nova.scheduler.client.report [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Deleted allocations for instance 30d82c70-1401-4a1a-a88b-f798f8fbf96a [ 797.578034] env[62000]: INFO nova.scheduler.client.report [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Deleted allocations for instance 49f6c198-13b7-4c07-81d8-c010e7b0598e [ 797.605471] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882217, 'name': ReconfigVM_Task, 'duration_secs': 0.301544} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.606322] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfigured VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 797.606788] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 797.624919] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882218, 'name': RemoveSnapshot_Task} progress is 65%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.673804] env[62000]: DEBUG oslo_concurrency.lockutils [None req-41de74ab-d910-4fdf-aa78-d65a550bfb86 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.859s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.745037] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882219, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 797.876088] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d8469c-50d3-a14f-6d39-6a37ebe237c2, 'name': SearchDatastore_Task, 'duration_secs': 0.036706} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 797.876954] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cd69b5e-707d-4d9c-952a-447f103f8544 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.883359] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 797.883359] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52114575-9895-f768-bb52-d7057202a564" [ 797.883359] env[62000]: _type = "Task" [ 797.883359] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 797.892369] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52114575-9895-f768-bb52-d7057202a564, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.053297] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] There are 13 instances to clean {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 798.053564] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: a9783a2a-a7d3-4fbe-9590-d530ac24fa82] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 798.093079] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.093394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.093598] env[62000]: DEBUG nova.compute.manager [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 798.094057] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7056aa1b-17ed-40c5-bc9b-f98c5617c70b tempest-ServersAaction247Test-687453193 tempest-ServersAaction247Test-687453193-project-member] Lock "30d82c70-1401-4a1a-a88b-f798f8fbf96a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 35.660s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.096380] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28569519-4ada-4fe6-b0e9-c90b13defa2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.100827] env[62000]: DEBUG oslo_concurrency.lockutils [None req-edc4979e-0dd1-4771-8072-d63f8923a475 tempest-ServersTestManualDisk-1157050087 tempest-ServersTestManualDisk-1157050087-project-member] Lock "49f6c198-13b7-4c07-81d8-c010e7b0598e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 38.980s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.108044] env[62000]: DEBUG nova.compute.manager [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 798.108044] env[62000]: DEBUG nova.objects.instance [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'flavor' on Instance uuid ae1e8ebc-ddb8-4e95-847e-b9684e9161d4 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 798.114188] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754d2e88-9d21-4310-84c0-97827f4271d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.127547] env[62000]: DEBUG oslo_vmware.api [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882218, 'name': RemoveSnapshot_Task, 'duration_secs': 0.618819} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.143049] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 798.143467] env[62000]: INFO nova.compute.manager [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 15.58 seconds to snapshot the instance on the hypervisor. [ 798.147343] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a43942-5567-45db-a0a5-35a2072e083a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.166994] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 798.244656] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882219, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.689269} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.244946] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 798.245191] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 798.245458] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-27cd7f3d-908a-4725-9e8e-d8763dc43ca0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.257641] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 798.257641] env[62000]: value = "task-882220" [ 798.257641] env[62000]: _type = "Task" [ 798.257641] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.266264] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882220, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.394875] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52114575-9895-f768-bb52-d7057202a564, 'name': SearchDatastore_Task, 'duration_secs': 0.025137} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.395195] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 798.395465] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 798.395732] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1eda0ba7-4a94-41d7-8201-dcea523bd41b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.405086] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 798.405086] env[62000]: value = "task-882222" [ 798.405086] env[62000]: _type = "Task" [ 798.405086] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.414866] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882222, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.560439] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: af49714d-8e50-4159-96a5-cf8f70580471] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 798.614732] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 798.614732] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cd7284ba-b4f6-4cbb-bb1c-9b72fcc7c752 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.628115] env[62000]: DEBUG oslo_vmware.api [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 798.628115] env[62000]: value = "task-882223" [ 798.628115] env[62000]: _type = "Task" [ 798.628115] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.641674] env[62000]: DEBUG oslo_vmware.api [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882223, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.701591] env[62000]: DEBUG nova.compute.manager [None req-b0d5f229-3888-4dcb-b3b3-439c070dad57 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Found 2 images (rotation: 2) {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 798.727329] env[62000]: DEBUG nova.network.neutron [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Port ddd21181-32c0-4898-bd09-7689d6976198 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 798.769378] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882220, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.181717} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 798.770151] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 798.774730] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bb16671-3f7e-4627-bb30-166f8a81cca0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.799341] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Reconfiguring VM instance instance-0000003c to attach disk [datastore2] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 798.804303] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-17d9f4c1-22c7-4ac2-9de3-a7023f31653d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.831306] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 798.831306] env[62000]: value = "task-882224" [ 798.831306] env[62000]: _type = "Task" [ 798.831306] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 798.851723] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882224, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.917937] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882222, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 798.988033] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82a44a8-b88d-4ff2-ad39-e7841e2881a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.996276] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-193280eb-8bee-4a3c-a424-0b9d830c604c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.036072] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e499084-28ee-420a-90f2-4a01f5b15fb0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.048042] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21017eec-c893-4688-9511-92913941ddf7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.068590] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 4e617bb5-84e4-4505-99e3-61289826f511] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 799.068844] env[62000]: DEBUG nova.compute.provider_tree [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.070894] env[62000]: DEBUG nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Successfully updated port: 656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 799.141455] env[62000]: DEBUG oslo_vmware.api [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882223, 'name': PowerOffVM_Task, 'duration_secs': 0.245692} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.141787] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 799.141919] env[62000]: DEBUG nova.compute.manager [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.142797] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c8d086d-d506-4817-a05d-627faa8eb94e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.343030] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882224, 'name': ReconfigVM_Task, 'duration_secs': 0.317775} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.343030] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Reconfigured VM instance instance-0000003c to attach disk [datastore2] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 799.343030] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-00f1d1ef-5617-4615-adfc-282083b2abe3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.351363] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 799.351363] env[62000]: value = "task-882225" [ 799.351363] env[62000]: _type = "Task" [ 799.351363] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.362015] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882225, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.418719] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882222, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.525486} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.418719] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 799.418719] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 799.418719] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64cfc58c-9b18-4199-8195-7ad0d4ec1ea3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.429234] env[62000]: DEBUG nova.compute.manager [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-plugged-656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 799.429511] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.429882] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.430580] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.430987] env[62000]: DEBUG nova.compute.manager [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] No waiting events found dispatching network-vif-plugged-656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 799.431344] env[62000]: WARNING nova.compute.manager [req-dc9bccab-5c04-4d9d-bf23-81f4976dce5b req-ca5fcf0f-6984-4364-b25b-237042aee0c9 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received unexpected event network-vif-plugged-656e8902-9da0-40c9-889a-8d1aab0f9266 for instance with vm_state active and task_state None. [ 799.434230] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 799.434230] env[62000]: value = "task-882226" [ 799.434230] env[62000]: _type = "Task" [ 799.434230] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.445665] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882226, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.576416] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 596af5ab-1791-42ce-93d2-3e4f0a47dfa2] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 799.577861] env[62000]: DEBUG nova.scheduler.client.report [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.581492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.581766] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.581864] env[62000]: DEBUG nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.609780] env[62000]: DEBUG nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 799.610877] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fcef03b-12ca-4b93-9b32-7c88b19cf5c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.655886] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2114ff0c-be0e-4431-b989-9e0b59c2cfc8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.562s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.752364] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.752561] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.752736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.861330] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882225, 'name': Rename_Task, 'duration_secs': 0.171988} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.861820] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 799.862224] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59de5491-24a4-4669-ad08-41201a651f07 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.870381] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 799.870381] env[62000]: value = "task-882227" [ 799.870381] env[62000]: _type = "Task" [ 799.870381] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.884850] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882227, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 799.945134] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882226, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 799.945435] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 799.946242] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d380ec-58a1-4bcc-bbee-13d912c9b53b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.972075] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 799.972075] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c03472fb-9b24-41bf-8c02-8fa2eaf7aadc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.995366] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 799.995366] env[62000]: value = "task-882228" [ 799.995366] env[62000]: _type = "Task" [ 799.995366] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.004115] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882228, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.084419] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 64ff4a7b-ec89-48cd-8fb6-124e0726d6f0] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 800.088249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.088918] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.096149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 24.545s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.096149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.098963] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.436s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.099319] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.102539] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.290s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.102839] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.105417] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.003s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.105668] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.110285] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.794s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.110425] env[62000]: INFO nova.compute.claims [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.124767] env[62000]: INFO nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] instance snapshotting [ 800.125144] env[62000]: DEBUG nova.objects.instance [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.162751] env[62000]: INFO nova.scheduler.client.report [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Deleted allocations for instance df525129-0ccb-4863-8a22-dd3e5a1aa2b5 [ 800.169659] env[62000]: INFO nova.scheduler.client.report [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Deleted allocations for instance 334fd514-c7b5-402f-935e-1d95f9b5dbc8 [ 800.171563] env[62000]: WARNING nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] e21c3795-1c6b-42ef-af81-e113912fa80d already exists in list: networks containing: ['e21c3795-1c6b-42ef-af81-e113912fa80d']. ignoring it [ 800.183842] env[62000]: INFO nova.scheduler.client.report [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Deleted allocations for instance 2b0c7354-1622-4318-ba09-6e3214eed4db [ 800.194816] env[62000]: INFO nova.scheduler.client.report [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Deleted allocations for instance 210af329-4cdb-4c3f-9e82-e72a2ea79421 [ 800.383872] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882227, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.508091] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882228, 'name': ReconfigVM_Task, 'duration_secs': 0.299822} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.509552] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Reconfigured VM instance instance-0000003a to attach disk [datastore2] ae343199-1006-473a-a47f-7983835bd60a/ae343199-1006-473a-a47f-7983835bd60a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 800.513087] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-876edf50-e5d6-4dc3-a589-36af0ac73ba8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.522025] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 800.522025] env[62000]: value = "task-882230" [ 800.522025] env[62000]: _type = "Task" [ 800.522025] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.532764] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882230, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.593750] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 72688992-333f-459d-9d05-f7c728961a6d] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 800.616669] env[62000]: DEBUG nova.compute.utils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.620170] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.620457] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 800.634833] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940f825d-c9c5-47ae-968a-a4c59835d7f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.655351] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17eed7d-116c-4d82-a9b5-5846705cfdb1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.681121] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7183cf6d-663a-4b39-b35e-1d247d13fa07 tempest-InstanceActionsNegativeTestJSON-1759008876 tempest-InstanceActionsNegativeTestJSON-1759008876-project-member] Lock "df525129-0ccb-4863-8a22-dd3e5a1aa2b5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 29.701s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.685671] env[62000]: DEBUG oslo_concurrency.lockutils [None req-286342b7-7aa7-42d6-82b5-1612d92df4de tempest-ServersTestBootFromVolume-116970681 tempest-ServersTestBootFromVolume-116970681-project-member] Lock "334fd514-c7b5-402f-935e-1d95f9b5dbc8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.534s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.693379] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ebc48f88-9b0e-4d66-a5c3-16f72d47359f tempest-ServerGroupTestJSON-600664980 tempest-ServerGroupTestJSON-600664980-project-member] Lock "2b0c7354-1622-4318-ba09-6e3214eed4db" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.154s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.705576] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b0c2287a-e078-4ef2-873e-1531057d8388 tempest-ServersV294TestFqdnHostnames-1502204228 tempest-ServersV294TestFqdnHostnames-1502204228-project-member] Lock "210af329-4cdb-4c3f-9e82-e72a2ea79421" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.453s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.711462] env[62000]: DEBUG nova.policy [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '041d24bce7f14a24b84fa108f731283b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c93358b7815d4b5ea3b188965c276329', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 800.845275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.845275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.845275] env[62000]: DEBUG nova.network.neutron [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 800.847272] env[62000]: DEBUG nova.network.neutron [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "656e8902-9da0-40c9-889a-8d1aab0f9266", "address": "fa:16:3e:24:53:f9", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap656e8902-9d", "ovs_interfaceid": "656e8902-9da0-40c9-889a-8d1aab0f9266", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.884534] env[62000]: DEBUG oslo_vmware.api [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882227, 'name': PowerOnVM_Task, 'duration_secs': 0.581077} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.884683] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 800.884952] env[62000]: INFO nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Took 5.87 seconds to spawn the instance on the hypervisor. [ 800.885177] env[62000]: DEBUG nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 800.887906] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e1481a-fa6c-4a63-b567-5caec6924f59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.898732] env[62000]: WARNING oslo_messaging._drivers.amqpdriver [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 801.035818] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882230, 'name': Rename_Task, 'duration_secs': 0.210128} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.035818] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 801.035818] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-21f35500-4c58-482e-818f-75a137e0cf9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.045695] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 801.045695] env[62000]: value = "task-882231" [ 801.045695] env[62000]: _type = "Task" [ 801.045695] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.061444] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882231, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.098203] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 415cc4a5-7610-4678-971d-cd00a0e8b54d] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 801.127119] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.166770] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 801.167122] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d051534c-76c6-42a3-a27f-c3a04b26f684 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.177528] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 801.177528] env[62000]: value = "task-882232" [ 801.177528] env[62000]: _type = "Task" [ 801.177528] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.194665] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882232, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.354486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.355211] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.355454] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.356703] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb99d64-b756-4186-af47-191d126c7306 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.381364] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 801.381695] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 801.381896] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 801.382200] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 801.383449] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 801.383701] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 801.383991] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 801.384279] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 801.384801] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 801.385046] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 801.385285] env[62000]: DEBUG nova.virt.hardware [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 801.393126] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfiguring VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 801.397108] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4cce9d04-2371-47f9-b9e4-fd6e0b409f91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.430244] env[62000]: INFO nova.compute.manager [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Took 41.21 seconds to build instance. [ 801.433421] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 801.433421] env[62000]: value = "task-882233" [ 801.433421] env[62000]: _type = "Task" [ 801.433421] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.447728] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.470221] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35cb4ada-7bff-4116-8344-0af4edbae48e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.479605] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df5ebe3b-3378-40fb-a41a-ebfee248520f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.521221] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Successfully created port: e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.524774] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-065eed2a-a461-4666-8d40-845121959e31 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.536677] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27ad70bf-cace-47c7-b9d6-603ebb6a2bc7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.558029] env[62000]: DEBUG nova.compute.provider_tree [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.571567] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882231, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.600216] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5e80a5c8-030c-4ad8-90c4-26136fa39d71] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 801.689166] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882232, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.886323] env[62000]: DEBUG nova.network.neutron [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.923240] env[62000]: DEBUG nova.compute.manager [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 801.924156] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d04edf59-c6c7-4fbc-87c8-f38ca0faf768 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.934119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54b1317c-d3b2-4cf1-92f0-249107309cd9 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.726s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.945913] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.065834] env[62000]: DEBUG nova.scheduler.client.report [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.073383] env[62000]: DEBUG oslo_vmware.api [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882231, 'name': PowerOnVM_Task, 'duration_secs': 0.546804} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.074625] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 802.074816] env[62000]: DEBUG nova.compute.manager [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 802.075712] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb83f6c0-d094-4b1d-9428-f1a2e3a071f3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.103777] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 1d9b67b5-c0cf-41ff-a838-79caf8789609] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 802.136445] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.167547] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.167863] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.167961] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.168488] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.168488] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.168488] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.168682] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.168952] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.169021] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.169209] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.169395] env[62000]: DEBUG nova.virt.hardware [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.170570] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8438ae-618c-4127-abe9-dde4c5304c8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.183103] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf05cc65-9fc3-4c95-a6a2-cc567c9679b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.195501] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882232, 'name': CreateSnapshot_Task, 'duration_secs': 0.917069} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.204218] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 802.205799] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-695eb04b-4ada-4e3c-8939-e92bb1d7679c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.388651] env[62000]: DEBUG oslo_concurrency.lockutils [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.442731] env[62000]: INFO nova.compute.manager [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] instance snapshotting [ 802.443135] env[62000]: WARNING nova.compute.manager [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] trying to snapshot a non-running instance: (state: 4 expected: 1) [ 802.449321] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-425bac6b-3e6d-4e46-a4bf-e124604662dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.457221] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.476152] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a077903-ba29-46a7-b6d1-a0f217ad0f18 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.577249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.469s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.578051] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.580239] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.573s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.580566] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.582715] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.275s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.582913] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.585302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.537s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.589149] env[62000]: INFO nova.compute.claims [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.600431] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.607606] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 762e429f-f690-43c4-95eb-877caf1cdad7] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 802.635341] env[62000]: INFO nova.scheduler.client.report [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Deleted allocations for instance 4c49fdd0-2485-4791-9349-a79a8663bbc0 [ 802.640602] env[62000]: INFO nova.scheduler.client.report [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Deleted allocations for instance 79a0d6f0-7dca-4c5d-8656-91bf4d957e78 [ 802.725129] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 802.725592] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-8e1c6e4e-9003-4b6a-998a-01a4db1ef859 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.737357] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 802.737357] env[62000]: value = "task-882235" [ 802.737357] env[62000]: _type = "Task" [ 802.737357] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.747374] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882235, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.784227] env[62000]: DEBUG nova.compute.manager [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-changed-656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 802.784743] env[62000]: DEBUG nova.compute.manager [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing instance network info cache due to event network-changed-656e8902-9da0-40c9-889a-8d1aab0f9266. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 802.787278] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.787278] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.787278] env[62000]: DEBUG nova.network.neutron [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing network info cache for port 656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 802.918793] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767dbb46-cc90-4dd3-93bd-dc4b5d7a1351 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.943627] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24a5d9f-3225-4da6-b585-bad62e7575b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.953235] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 802.967095] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.991246] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 802.991246] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1565100e-8b72-4e38-b83a-d3df0af48440 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.006343] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 803.006343] env[62000]: value = "task-882236" [ 803.006343] env[62000]: _type = "Task" [ 803.006343] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.043620] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882236, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.096327] env[62000]: DEBUG nova.compute.utils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.097924] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.102065] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 803.114141] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 1298037f-9cb7-4e2b-b70f-2ab24efe9b91] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 803.155366] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8d0797a4-10f0-4fad-851c-f075186d91b3 tempest-ServersTestMultiNic-398619037 tempest-ServersTestMultiNic-398619037-project-member] Lock "79a0d6f0-7dca-4c5d-8656-91bf4d957e78" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 18.302s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.156914] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be2a3412-f712-4f15-9ab5-693e0e19bb9b tempest-ServerTagsTestJSON-1154676420 tempest-ServerTagsTestJSON-1154676420-project-member] Lock "4c49fdd0-2485-4791-9349-a79a8663bbc0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.501s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.253684] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882235, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.426560] env[62000]: DEBUG nova.policy [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9277350d53c24970a49d71aff09fe192', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '817c2729a04843518a4333c6203a29b7', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.457554] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.475926] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 803.475926] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c0b1af4a-df7f-46f1-9a7a-47d3114b7536 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.489458] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 803.489458] env[62000]: value = "task-882237" [ 803.489458] env[62000]: _type = "Task" [ 803.489458] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.507351] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882237, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.519037] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882236, 'name': CreateSnapshot_Task, 'duration_secs': 0.430774} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.519160] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 803.520377] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e39818-ffb4-4dc3-aa6e-ed9952b0f377 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.598131] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.619904] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 1419e5a8-a78f-44f1-bfc4-00fcf176bf35] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 803.708535] env[62000]: INFO nova.compute.manager [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Rebuilding instance [ 803.756814] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882235, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.775487] env[62000]: DEBUG nova.compute.manager [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 803.777632] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44bc3ba2-753a-4869-a876-0b8b51f4c1e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.896212] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4804e1f4-8ff0-48e3-86a7-bf37d68cf924 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.907757] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e34e06f-d9b8-41b6-a375-0143035e951c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.956183] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f31e1a9-3665-43ff-ae03-b20e603cba6e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.965562] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.968857] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-373d2595-ac82-4ac6-bf7b-1684364babd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.984868] env[62000]: DEBUG nova.compute.provider_tree [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.003180] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882237, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.030397] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Successfully created port: d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 804.039896] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 804.040632] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-a7a83fbd-8c2f-46d2-a8db-24a304320fa7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.054190] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 804.054190] env[62000]: value = "task-882238" [ 804.054190] env[62000]: _type = "Task" [ 804.054190] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.069199] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.124445] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 15409fa7-254c-435e-8080-46d3f65b2d46] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 804.259497] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882235, 'name': CloneVM_Task, 'duration_secs': 1.165205} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.259871] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Created linked-clone VM from snapshot [ 804.260767] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46ee6a82-6608-4a43-9b2b-75e09a983738 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.273577] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploading image e83c8ab9-d1dc-497c-bb7e-7cd42feab5a7 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 804.298807] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 804.299604] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0f71d275-c887-440d-9583-ff39a3a635ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.313597] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 804.313597] env[62000]: value = "task-882239" [ 804.313597] env[62000]: _type = "Task" [ 804.313597] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.316529] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 804.316529] env[62000]: value = "vm-201544" [ 804.316529] env[62000]: _type = "VirtualMachine" [ 804.316529] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 804.316992] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-7b552978-7b91-41e8-a8ad-1fc9ebaeaa89 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.333630] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882239, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.335653] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease: (returnval){ [ 804.335653] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520b4caa-baf8-a71d-f0b0-3a4b1f7217a8" [ 804.335653] env[62000]: _type = "HttpNfcLease" [ 804.335653] env[62000]: } obtained for exporting VM: (result){ [ 804.335653] env[62000]: value = "vm-201544" [ 804.335653] env[62000]: _type = "VirtualMachine" [ 804.335653] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 804.336103] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the lease: (returnval){ [ 804.336103] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520b4caa-baf8-a71d-f0b0-3a4b1f7217a8" [ 804.336103] env[62000]: _type = "HttpNfcLease" [ 804.336103] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 804.347711] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 804.347711] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520b4caa-baf8-a71d-f0b0-3a4b1f7217a8" [ 804.347711] env[62000]: _type = "HttpNfcLease" [ 804.347711] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 804.468500] env[62000]: DEBUG oslo_vmware.api [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882233, 'name': ReconfigVM_Task, 'duration_secs': 2.90124} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.469610] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.470284] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfigured VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 804.488806] env[62000]: DEBUG nova.scheduler.client.report [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.516079] env[62000]: DEBUG oslo_vmware.api [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882237, 'name': PowerOnVM_Task, 'duration_secs': 0.790826} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.516269] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 804.516829] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-453cebb2-4476-43a1-8824-93cc8bcdcf5a tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance 'f3a6527c-777e-4a07-9482-598de15d4eb3' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 804.565784] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.610225] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.631082] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 804.631336] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Cleaning up deleted instances with incomplete migration {{(pid=62000) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.657061] env[62000]: DEBUG nova.virt.hardware [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.657061] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f85ce5b-7f10-402a-8886-f42a78aa0b8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.669347] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7a2612c-1a31-464b-ae24-11ef873aebd7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.779922] env[62000]: DEBUG nova.network.neutron [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updated VIF entry in instance network info cache for port 656e8902-9da0-40c9-889a-8d1aab0f9266. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 804.780677] env[62000]: DEBUG nova.network.neutron [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "656e8902-9da0-40c9-889a-8d1aab0f9266", "address": "fa:16:3e:24:53:f9", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap656e8902-9d", "ovs_interfaceid": "656e8902-9da0-40c9-889a-8d1aab0f9266", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.831978] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882239, 'name': PowerOffVM_Task, 'duration_secs': 0.207373} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.834461] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 804.835825] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 804.836080] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80ecc7e-51c1-4fac-8a48-7f46efa3ac91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.850762] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 804.850762] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520b4caa-baf8-a71d-f0b0-3a4b1f7217a8" [ 804.850762] env[62000]: _type = "HttpNfcLease" [ 804.850762] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 804.853978] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 804.853978] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520b4caa-baf8-a71d-f0b0-3a4b1f7217a8" [ 804.853978] env[62000]: _type = "HttpNfcLease" [ 804.853978] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 804.856019] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 804.856019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b744a6be-d767-4281-bebe-3a6030860b04 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.858629] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ddcd3719-93e0-4fc6-9527-5ce347720138 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.869410] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 804.869739] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 804.935499] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 804.935743] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 804.935960] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Deleting the datastore file [datastore2] 153b86d9-dc2e-463d-87fd-155ec23e2abb {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 804.937020] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-949bf91a-486b-4b15-9147-65e070126303 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.948891] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 804.948891] env[62000]: value = "task-882242" [ 804.948891] env[62000]: _type = "Task" [ 804.948891] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 804.960643] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882242, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 804.976824] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5ec39f80-eccb-4ab9-a807-4a2fe90f3a7e tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.720s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.983677] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-02a335ca-7fdc-41ac-829b-2bce0462e40a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.005532] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.005532] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 805.006664] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.035s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.007078] env[62000]: DEBUG nova.objects.instance [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lazy-loading 'resources' on Instance uuid 296d5fea-e9d2-49e3-bc29-8e4bf3782535 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 805.069670] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.120795] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Successfully updated port: e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 805.137906] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 805.287723] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ad88af3-6e1c-419e-aeaf-f0e05f39cafc req-e4bfad31-16ba-4f0a-83ca-f146dd2dc6f7 service nova] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.465076] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882242, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.143461} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 805.465320] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 805.466682] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 805.466682] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 805.513408] env[62000]: DEBUG nova.compute.utils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.523614] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.523822] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 805.569125] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 805.627731] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.627976] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.628613] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.734831] env[62000]: DEBUG nova.policy [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79730e91610c4c598019b122b4f41418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6106a70abdad4111977eebbd82434337', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.902701] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd686967-e650-4312-9eda-09df3f0625d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.914913] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b02a1023-a204-41e4-b610-9b10975392a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.978827] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35494393-e9bd-4d12-a51b-32cf41871ac9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.989156] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9265ce02-9b69-4131-b70d-21b582bf6e30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.007758] env[62000]: DEBUG nova.compute.provider_tree [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.026529] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 806.083177] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.250677] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.315592] env[62000]: DEBUG nova.compute.manager [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Received event network-vif-plugged-e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 806.319610] env[62000]: DEBUG oslo_concurrency.lockutils [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] Acquiring lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.319949] env[62000]: DEBUG oslo_concurrency.lockutils [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 806.320158] env[62000]: DEBUG oslo_concurrency.lockutils [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.320351] env[62000]: DEBUG nova.compute.manager [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] No waiting events found dispatching network-vif-plugged-e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 806.320536] env[62000]: WARNING nova.compute.manager [req-e03f6469-6ec0-4d87-b8ca-45685a92a5c0 req-94fd1b5f-e1b1-455c-bd9b-554066e46d9e service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Received unexpected event network-vif-plugged-e3405b94-e374-4e74-8dc3-f1bb44464833 for instance with vm_state building and task_state spawning. [ 806.512901] env[62000]: DEBUG nova.scheduler.client.report [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 806.520327] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.523896] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.525403] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.525855] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.526237] env[62000]: DEBUG nova.virt.hardware [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.527888] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682e9692-de77-4d5f-b3b5-e3d38bb9342b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.545268] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66efaa56-7127-4c07-98a9-7eab98f7ea60 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.563778] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 806.570517] env[62000]: DEBUG oslo.service.loopingcall [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 806.574110] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 806.578431] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-10a254b7-f706-45a3-b9c8-5d9e0e85c3eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.601618] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882238, 'name': CloneVM_Task, 'duration_secs': 2.069709} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.603290] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Created linked-clone VM from snapshot [ 806.603683] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 806.603683] env[62000]: value = "task-882243" [ 806.603683] env[62000]: _type = "Task" [ 806.603683] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.604664] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f53bede8-4bff-4972-bbcc-0044c5bc57e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.618563] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882243, 'name': CreateVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.623288] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Uploading image 53a44733-1c2c-4740-b053-0d40c24bf75b {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 806.649806] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 806.649806] env[62000]: value = "vm-201546" [ 806.649806] env[62000]: _type = "VirtualMachine" [ 806.649806] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 806.650865] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-18d6eacc-530c-463a-8e75-4b2e924c02f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.662290] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease: (returnval){ [ 806.662290] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52521dd2-55fe-f7e8-f708-509788821c14" [ 806.662290] env[62000]: _type = "HttpNfcLease" [ 806.662290] env[62000]: } obtained for exporting VM: (result){ [ 806.662290] env[62000]: value = "vm-201546" [ 806.662290] env[62000]: _type = "VirtualMachine" [ 806.662290] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 806.662290] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the lease: (returnval){ [ 806.662290] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52521dd2-55fe-f7e8-f708-509788821c14" [ 806.662290] env[62000]: _type = "HttpNfcLease" [ 806.662290] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 806.672053] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 806.672053] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52521dd2-55fe-f7e8-f708-509788821c14" [ 806.672053] env[62000]: _type = "HttpNfcLease" [ 806.672053] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 807.019159] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.013s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.022349] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.575s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.024300] env[62000]: INFO nova.compute.claims [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.040884] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 807.050453] env[62000]: INFO nova.scheduler.client.report [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Deleted allocations for instance 296d5fea-e9d2-49e3-bc29-8e4bf3782535 [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 807.078018] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 807.078325] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 807.078325] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 807.078537] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 807.078707] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 807.078886] env[62000]: DEBUG nova.virt.hardware [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 807.080766] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef7b7b4c-f860-4c9a-88ff-df646e448e0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.084431] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Successfully updated port: d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 807.096787] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f139d1e1-69d1-43af-bf01-03d432d2a7a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.129269] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882243, 'name': CreateVM_Task, 'duration_secs': 0.384659} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.129269] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 807.130404] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.130521] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.131514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 807.132683] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bfc13f5-0623-42e9-9631-fcccd8b295e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.142495] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 807.142495] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5246776e-a302-b896-a238-2dd6255c3632" [ 807.142495] env[62000]: _type = "Task" [ 807.142495] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.153971] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5246776e-a302-b896-a238-2dd6255c3632, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.170374] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 807.170374] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52521dd2-55fe-f7e8-f708-509788821c14" [ 807.170374] env[62000]: _type = "HttpNfcLease" [ 807.170374] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 807.170836] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 807.170836] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52521dd2-55fe-f7e8-f708-509788821c14" [ 807.170836] env[62000]: _type = "HttpNfcLease" [ 807.170836] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 807.171515] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb248513-839b-4bc2-89e9-b92be8ff0d23 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.181707] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 807.182146] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 807.249200] env[62000]: DEBUG nova.network.neutron [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Updating instance_info_cache with network_info: [{"id": "e3405b94-e374-4e74-8dc3-f1bb44464833", "address": "fa:16:3e:23:45:15", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3405b94-e3", "ovs_interfaceid": "e3405b94-e374-4e74-8dc3-f1bb44464833", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.317464] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-cbe9f9a2-e07a-4e85-b2d1-ae6e54474826 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.383345] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Successfully created port: 0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 807.562061] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4622319-7dde-4233-a9a7-ff42c3edffcd tempest-TenantUsagesTestJSON-391045992 tempest-TenantUsagesTestJSON-391045992-project-member] Lock "296d5fea-e9d2-49e3-bc29-8e4bf3782535" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 15.518s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.590590] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.590590] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquired lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.590590] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.654401] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5246776e-a302-b896-a238-2dd6255c3632, 'name': SearchDatastore_Task, 'duration_secs': 0.012737} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.654789] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.655062] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 807.655371] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.655532] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.655733] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 807.656042] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6068914b-969f-4612-b671-30ee6d6d249e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.668487] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 807.668804] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 807.670242] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfdcc80b-ff16-4f5d-a94c-e86edeff6ec1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.679201] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 807.679201] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526657bb-02f0-a16d-00f4-9d4b591568ed" [ 807.679201] env[62000]: _type = "Task" [ 807.679201] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.688641] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526657bb-02f0-a16d-00f4-9d4b591568ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 807.755457] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 807.755457] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Instance network_info: |[{"id": "e3405b94-e374-4e74-8dc3-f1bb44464833", "address": "fa:16:3e:23:45:15", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3405b94-e3", "ovs_interfaceid": "e3405b94-e374-4e74-8dc3-f1bb44464833", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 807.755457] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:23:45:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '360308f4-9d0a-4ec2-8bcf-44891f452847', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e3405b94-e374-4e74-8dc3-f1bb44464833', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 807.765334] env[62000]: DEBUG oslo.service.loopingcall [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 807.765879] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 807.766284] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c6c8bfaa-3dfc-4375-9fee-f1aeb8367c2f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.793241] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 807.793241] env[62000]: value = "task-882245" [ 807.793241] env[62000]: _type = "Task" [ 807.793241] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 807.803874] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882245, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.200359] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526657bb-02f0-a16d-00f4-9d4b591568ed, 'name': SearchDatastore_Task, 'duration_secs': 0.016707} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.210929] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-209dc128-f78e-45d0-a51b-f9bc6d3e6ef8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.222438] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 808.222438] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522d7c2f-30cc-bbeb-c067-514efc54ced3" [ 808.222438] env[62000]: _type = "Task" [ 808.222438] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.241138] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522d7c2f-30cc-bbeb-c067-514efc54ced3, 'name': SearchDatastore_Task, 'duration_secs': 0.014568} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.241450] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.241757] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 808.242117] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33e1ec6e-82cc-4e99-b022-7fdf27a62f3a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.252702] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 808.252702] env[62000]: value = "task-882246" [ 808.252702] env[62000]: _type = "Task" [ 808.252702] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.260944] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.269227] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882246, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.310963] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882245, 'name': CreateVM_Task, 'duration_secs': 0.461613} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.315435] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 808.316591] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.317201] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.317435] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.318035] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86ecc15d-9ceb-40dc-996a-727f2993f8f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.326041] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 808.326041] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c36ad5-7e8f-b109-b334-9600b2b24444" [ 808.326041] env[62000]: _type = "Task" [ 808.326041] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.343781] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c36ad5-7e8f-b109-b334-9600b2b24444, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.398205] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16292ca0-21d7-4b02-b61c-245237f12599 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.405923] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f60ea23e-31b9-4694-b406-272a4585b44e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.447407] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c74f3ca-d887-41a5-9072-a87dbe1cea22 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.461730] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c6e61f-57f6-466c-be3f-685654b1f135 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.477982] env[62000]: DEBUG nova.compute.provider_tree [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.769898] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882246, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.811525] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.811800] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.845377] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c36ad5-7e8f-b109-b334-9600b2b24444, 'name': SearchDatastore_Task, 'duration_secs': 0.016564} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.845699] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.845962] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 808.846220] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.846366] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.846561] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 808.846873] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-59656a4b-5ff5-4890-9fd5-b9af95be7b44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.867382] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 808.867815] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 808.869750] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8bde5ba2-af58-4e3a-856c-d50e32a2d090 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.880039] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 808.880039] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc1ae-329e-5748-f2e3-05694592248e" [ 808.880039] env[62000]: _type = "Task" [ 808.880039] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.892759] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc1ae-329e-5748-f2e3-05694592248e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.931030] env[62000]: DEBUG nova.network.neutron [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Updating instance_info_cache with network_info: [{"id": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "address": "fa:16:3e:b6:d0:9f", "network": {"id": "2d839574-e381-4fdf-970c-69737dbc35dc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1373885955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "817c2729a04843518a4333c6203a29b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f9be6786-e9a7-4138-b7b5-b7696f6cb1e1", "external-id": "nsx-vlan-transportzone-626", "segmentation_id": 626, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8f74268-0a", "ovs_interfaceid": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.983305] env[62000]: DEBUG nova.compute.manager [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Received event network-vif-plugged-d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 808.986198] env[62000]: DEBUG oslo_concurrency.lockutils [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] Acquiring lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.986198] env[62000]: DEBUG oslo_concurrency.lockutils [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.986995] env[62000]: DEBUG oslo_concurrency.lockutils [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 808.987661] env[62000]: DEBUG nova.compute.manager [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] No waiting events found dispatching network-vif-plugged-d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 808.987661] env[62000]: WARNING nova.compute.manager [req-45443eb7-c990-4cb4-abd3-c21d322655a4 req-c1d814a8-7e20-490d-a542-e8e8e58d7f99 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Received unexpected event network-vif-plugged-d8f74268-0a99-4e0d-8042-52662b7e2c21 for instance with vm_state building and task_state spawning. [ 808.989385] env[62000]: DEBUG nova.scheduler.client.report [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.265561] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882246, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.70855} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.265868] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 809.266124] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 809.266478] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0206b0f4-b909-4d09-8607-0474ee999f91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.278599] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 809.278599] env[62000]: value = "task-882247" [ 809.278599] env[62000]: _type = "Task" [ 809.278599] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.291445] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882247, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.321077] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 809.400610] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc1ae-329e-5748-f2e3-05694592248e, 'name': SearchDatastore_Task, 'duration_secs': 0.053243} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.401214] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-026f5a93-2026-4ea8-ad90-7102b07cabbd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.410032] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 809.410032] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52083542-3669-47c5-1975-8398d5256d3e" [ 809.410032] env[62000]: _type = "Task" [ 809.410032] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.420419] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52083542-3669-47c5-1975-8398d5256d3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.436215] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Releasing lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.436628] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Instance network_info: |[{"id": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "address": "fa:16:3e:b6:d0:9f", "network": {"id": "2d839574-e381-4fdf-970c-69737dbc35dc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1373885955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "817c2729a04843518a4333c6203a29b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f9be6786-e9a7-4138-b7b5-b7696f6cb1e1", "external-id": "nsx-vlan-transportzone-626", "segmentation_id": 626, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8f74268-0a", "ovs_interfaceid": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 809.437137] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:d0:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f9be6786-e9a7-4138-b7b5-b7696f6cb1e1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8f74268-0a99-4e0d-8042-52662b7e2c21', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 809.448021] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Creating folder: Project (817c2729a04843518a4333c6203a29b7). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.448021] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c5bb2577-0926-4a6a-b14b-fb05b1ef6f51 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.462300] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Created folder: Project (817c2729a04843518a4333c6203a29b7) in parent group-v201431. [ 809.462557] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Creating folder: Instances. Parent ref: group-v201549. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 809.466022] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3aed6076-1770-4bca-9abe-df09b007cef9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.473428] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "48d29490-c174-4702-a449-3bf10bc9207c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.473769] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.476893] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Created folder: Instances in parent group-v201549. [ 809.477230] env[62000]: DEBUG oslo.service.loopingcall [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 809.477718] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 809.477968] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-55bf11d2-55ff-436d-bfa5-2885366b724c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.494927] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.473s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.495569] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 809.498951] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.899s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.499233] env[62000]: DEBUG nova.objects.instance [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 809.508140] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 809.508140] env[62000]: value = "task-882250" [ 809.508140] env[62000]: _type = "Task" [ 809.508140] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.518350] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882250, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.572766] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "ae343199-1006-473a-a47f-7983835bd60a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.573143] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.573400] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "ae343199-1006-473a-a47f-7983835bd60a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.573608] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.573803] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.576453] env[62000]: INFO nova.compute.manager [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Terminating instance [ 809.579630] env[62000]: DEBUG nova.compute.manager [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 809.579630] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 809.580197] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe068cc-7a20-435c-9806-d1ce46666801 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.591469] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 809.591469] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d557459e-f7cf-4021-a892-d434f0e65617 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.598990] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 809.598990] env[62000]: value = "task-882251" [ 809.598990] env[62000]: _type = "Task" [ 809.598990] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.608754] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882251, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.796171] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882247, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077878} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.796171] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 809.797935] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff91c1f1-6c5e-482f-bad5-b54ebf9752ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.827166] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 809.830155] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ecb41998-4546-473c-80a4-b34b10d2ee99 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.861382] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 809.861382] env[62000]: value = "task-882252" [ 809.861382] env[62000]: _type = "Task" [ 809.861382] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.870201] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.874951] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882252, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.925454] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52083542-3669-47c5-1975-8398d5256d3e, 'name': SearchDatastore_Task, 'duration_secs': 0.015457} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.925986] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.929984] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 96107e36-bf3f-4ef0-8d8b-5c9601f4f514/96107e36-bf3f-4ef0-8d8b-5c9601f4f514.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 809.930451] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0c7e11da-832a-444c-b083-5abea8865501 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.940326] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 809.940326] env[62000]: value = "task-882253" [ 809.940326] env[62000]: _type = "Task" [ 809.940326] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.955661] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882253, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.979482] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 810.004541] env[62000]: DEBUG nova.compute.utils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 810.011664] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 810.011931] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 810.031907] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882250, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.095067] env[62000]: DEBUG nova.policy [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '674db9508719490ba7e22a3c5108837a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6742b3d48eb945fe8a11d0ee651abd1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 810.114793] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882251, 'name': PowerOffVM_Task, 'duration_secs': 0.344043} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.114793] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 810.114793] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 810.114793] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-28b7de6b-d214-4ae6-9770-30d35a05a1fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.280742] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 810.281021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 810.281226] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore2] ae343199-1006-473a-a47f-7983835bd60a {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 810.281504] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df71cf57-4a5d-4d6d-9e9a-4d28d8f159f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.291135] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 810.291135] env[62000]: value = "task-882255" [ 810.291135] env[62000]: _type = "Task" [ 810.291135] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.305082] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882255, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.377721] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882252, 'name': ReconfigVM_Task, 'duration_secs': 0.37767} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.380698] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Reconfigured VM instance instance-0000003c to attach disk [datastore1] 153b86d9-dc2e-463d-87fd-155ec23e2abb/153b86d9-dc2e-463d-87fd-155ec23e2abb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 810.380698] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a28a88e6-b92f-48a0-a855-0ab0f155da5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.393067] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 810.393067] env[62000]: value = "task-882256" [ 810.393067] env[62000]: _type = "Task" [ 810.393067] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.412228] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882256, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.453097] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882253, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.506746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.515293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a1e8cbf6-236d-4bb9-b70d-1e59eef4f557 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.515293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.645s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.516571] env[62000]: INFO nova.compute.claims [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 810.520810] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 810.539458] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882250, 'name': CreateVM_Task, 'duration_secs': 0.605314} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.542215] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 810.543133] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.545282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.545639] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 810.545923] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d1d58f1-e9a9-406f-992d-f173c99951fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.554189] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 810.554189] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64cdc-e99b-286b-ab43-21efb8b0b832" [ 810.554189] env[62000]: _type = "Task" [ 810.554189] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.569317] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64cdc-e99b-286b-ab43-21efb8b0b832, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.806081] env[62000]: DEBUG oslo_vmware.api [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882255, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.322623} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.806255] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 810.807265] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 810.807265] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 810.807449] env[62000]: INFO nova.compute.manager [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: ae343199-1006-473a-a47f-7983835bd60a] Took 1.23 seconds to destroy the instance on the hypervisor. [ 810.808640] env[62000]: DEBUG oslo.service.loopingcall [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 810.808640] env[62000]: DEBUG nova.compute.manager [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.808640] env[62000]: DEBUG nova.network.neutron [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.916474] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882256, 'name': Rename_Task, 'duration_secs': 0.23605} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.917095] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 810.917618] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f33346f6-b22b-47be-92ce-314448589a34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.929978] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 810.929978] env[62000]: value = "task-882257" [ 810.929978] env[62000]: _type = "Task" [ 810.929978] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.947959] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882257, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.963347] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882253, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.6491} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.963887] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 96107e36-bf3f-4ef0-8d8b-5c9601f4f514/96107e36-bf3f-4ef0-8d8b-5c9601f4f514.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 810.964329] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.964944] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dcdfde76-cdeb-473c-ac9b-da1679637864 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.974968] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 810.974968] env[62000]: value = "task-882258" [ 810.974968] env[62000]: _type = "Task" [ 810.974968] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.988152] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882258, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.028337] env[62000]: DEBUG nova.compute.manager [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Received event network-changed-e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 811.028337] env[62000]: DEBUG nova.compute.manager [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Refreshing instance network info cache due to event network-changed-e3405b94-e374-4e74-8dc3-f1bb44464833. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 811.028337] env[62000]: DEBUG oslo_concurrency.lockutils [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] Acquiring lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.028337] env[62000]: DEBUG oslo_concurrency.lockutils [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] Acquired lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.028337] env[62000]: DEBUG nova.network.neutron [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Refreshing network info cache for port e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 811.071151] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e64cdc-e99b-286b-ab43-21efb8b0b832, 'name': SearchDatastore_Task, 'duration_secs': 0.025487} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.071548] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 811.071852] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 811.072216] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.073652] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.073652] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 811.073652] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7b2b4829-3e69-48e7-91a1-0a7cd70ce33e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.086146] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 811.086486] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 811.088055] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2e951448-1b02-4d0c-a2d6-bad003660ad9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.098169] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 811.098169] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c2dde5-97f1-6f35-d99c-f11eb05c7f63" [ 811.098169] env[62000]: _type = "Task" [ 811.098169] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.111404] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c2dde5-97f1-6f35-d99c-f11eb05c7f63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.292451] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Successfully updated port: 0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 811.429155] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Successfully created port: 849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 811.445680] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882257, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.486158] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882258, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.109309} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.488272] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.488498] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69981ca5-ccac-4862-8a44-c037f8b8635b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.520052] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] 96107e36-bf3f-4ef0-8d8b-5c9601f4f514/96107e36-bf3f-4ef0-8d8b-5c9601f4f514.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.520810] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b8093b09-139d-43d0-a776-c72a5835fef4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.551333] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 811.557545] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 811.557545] env[62000]: value = "task-882259" [ 811.557545] env[62000]: _type = "Task" [ 811.557545] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.568529] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882259, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.611872] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c2dde5-97f1-6f35-d99c-f11eb05c7f63, 'name': SearchDatastore_Task, 'duration_secs': 0.016099} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.615226] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b5edcc5-ae7a-4f2c-8a2b-5291da60b23b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.625719] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 811.625719] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522507ed-e332-db6e-88f2-00ba0ad18465" [ 811.625719] env[62000]: _type = "Task" [ 811.625719] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.637115] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522507ed-e332-db6e-88f2-00ba0ad18465, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.796271] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.796369] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.797011] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.858144] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74a1d93-19e4-47a9-a98c-692e910e660b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.868440] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2f07c0-2244-4022-84cc-56e9a59ef811 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.913715] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638f2ff4-bb09-4575-81c9-50f2d61077fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.923897] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2397b359-69c9-4f7d-b909-4908409d299d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.930180] env[62000]: DEBUG nova.network.neutron [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.945202] env[62000]: DEBUG nova.compute.provider_tree [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 811.955039] env[62000]: DEBUG oslo_vmware.api [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882257, 'name': PowerOnVM_Task, 'duration_secs': 0.611873} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.955521] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 811.955794] env[62000]: DEBUG nova.compute.manager [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 811.957708] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffd2e5b0-a14a-4653-9fb1-efd22756df65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.992652] env[62000]: DEBUG nova.network.neutron [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Updated VIF entry in instance network info cache for port e3405b94-e374-4e74-8dc3-f1bb44464833. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 811.996250] env[62000]: DEBUG nova.network.neutron [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Updating instance_info_cache with network_info: [{"id": "e3405b94-e374-4e74-8dc3-f1bb44464833", "address": "fa:16:3e:23:45:15", "network": {"id": "f196ec2f-77c4-4ad3-bf77-afcffd6f0287", "bridge": "br-int", "label": "tempest-VolumesAdminNegativeTest-1787687502-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c93358b7815d4b5ea3b188965c276329", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "360308f4-9d0a-4ec2-8bcf-44891f452847", "external-id": "nsx-vlan-transportzone-383", "segmentation_id": 383, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape3405b94-e3", "ovs_interfaceid": "e3405b94-e374-4e74-8dc3-f1bb44464833", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.075418] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882259, 'name': ReconfigVM_Task, 'duration_secs': 0.413856} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.075987] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Reconfigured VM instance instance-0000003d to attach disk [datastore1] 96107e36-bf3f-4ef0-8d8b-5c9601f4f514/96107e36-bf3f-4ef0-8d8b-5c9601f4f514.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 812.077290] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f92ffa5-afec-42ba-8176-8a0ff5b774d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.087453] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 812.087453] env[62000]: value = "task-882260" [ 812.087453] env[62000]: _type = "Task" [ 812.087453] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.097459] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882260, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.142105] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522507ed-e332-db6e-88f2-00ba0ad18465, 'name': SearchDatastore_Task, 'duration_secs': 0.013725} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.142105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.143211] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 85f5f90a-9591-4393-9bef-2ad7e6a1a82f/85f5f90a-9591-4393-9bef-2ad7e6a1a82f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 812.143211] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9d73d95d-2649-4dce-a03e-ed41bb8625db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.153879] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 812.153879] env[62000]: value = "task-882261" [ 812.153879] env[62000]: _type = "Task" [ 812.153879] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.169773] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882261, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.247424] env[62000]: DEBUG nova.network.neutron [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Port ddd21181-32c0-4898-bd09-7689d6976198 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 812.247676] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.247837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.248028] env[62000]: DEBUG nova.network.neutron [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 812.381744] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.432430] env[62000]: INFO nova.compute.manager [-] [instance: ae343199-1006-473a-a47f-7983835bd60a] Took 1.62 seconds to deallocate network for instance. [ 812.448994] env[62000]: DEBUG nova.scheduler.client.report [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.488333] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.499397] env[62000]: DEBUG oslo_concurrency.lockutils [req-e195b6a9-9c9d-42ed-9802-ef8bd85668f2 req-4a39e993-4690-4bc8-95a9-90670963750d service nova] Releasing lock "refresh_cache-96107e36-bf3f-4ef0-8d8b-5c9601f4f514" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.601952] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882260, 'name': Rename_Task, 'duration_secs': 0.17614} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.603017] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 812.603548] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cfafc023-1aa7-487c-b1e8-dee0e63912a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.616856] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 812.616856] env[62000]: value = "task-882262" [ 812.616856] env[62000]: _type = "Task" [ 812.616856] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.633581] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882262, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.669012] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882261, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.770220] env[62000]: DEBUG nova.network.neutron [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Updating instance_info_cache with network_info: [{"id": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "address": "fa:16:3e:03:c1:8c", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cb92206-16", "ovs_interfaceid": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.939495] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.957825] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.444s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.959647] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 812.965439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.456s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.965439] env[62000]: INFO nova.compute.claims [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.127355] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882262, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.145345] env[62000]: DEBUG nova.network.neutron [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.168298] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882261, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591881} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.168658] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 85f5f90a-9591-4393-9bef-2ad7e6a1a82f/85f5f90a-9591-4393-9bef-2ad7e6a1a82f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 813.168901] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 813.169193] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a3587806-6768-4956-863f-bbcd2291b78f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.181591] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 813.181591] env[62000]: value = "task-882263" [ 813.181591] env[62000]: _type = "Task" [ 813.181591] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.193892] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882263, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.262324] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Successfully updated port: 849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 813.276964] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.277935] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance network_info: |[{"id": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "address": "fa:16:3e:03:c1:8c", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cb92206-16", "ovs_interfaceid": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 813.279175] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:c1:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0cb92206-16ee-49e1-8e70-fa3f39d678b4', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 813.288117] env[62000]: DEBUG oslo.service.loopingcall [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 813.288819] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 813.289117] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1b4e04a1-29ed-43e1-9eef-207c2620dec8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.314394] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 813.314394] env[62000]: value = "task-882264" [ 813.314394] env[62000]: _type = "Task" [ 813.314394] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.325464] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882264, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.479352] env[62000]: DEBUG nova.compute.utils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 813.479352] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 813.479352] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 813.592953] env[62000]: DEBUG nova.policy [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36534ad16b49477abe533ac313f2819a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7d4b4854b30648d8ac861f08fbdb6da9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 813.638875] env[62000]: DEBUG oslo_vmware.api [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882262, 'name': PowerOnVM_Task, 'duration_secs': 0.675952} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.641475] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 813.641819] env[62000]: INFO nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Took 11.51 seconds to spawn the instance on the hypervisor. [ 813.642151] env[62000]: DEBUG nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 813.643962] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbc487a-0531-4709-936e-5f14d8edfa86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.648603] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.694416] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882263, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.130297} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 813.694802] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 813.695694] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4ad2aca-6e68-4a65-8f82-918f1beef064 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.722578] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 85f5f90a-9591-4393-9bef-2ad7e6a1a82f/85f5f90a-9591-4393-9bef-2ad7e6a1a82f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 813.725694] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b0daa8e4-3033-40e1-ba09-93620de4cb73 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.751167] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 813.751167] env[62000]: value = "task-882265" [ 813.751167] env[62000]: _type = "Task" [ 813.751167] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 813.761398] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882265, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.770478] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.770739] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.770834] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.825761] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882264, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 813.984276] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 814.158492] env[62000]: DEBUG nova.compute.manager [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62000) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 814.158751] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.182978] env[62000]: INFO nova.compute.manager [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Took 43.04 seconds to build instance. [ 814.265022] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882265, 'name': ReconfigVM_Task, 'duration_secs': 0.507497} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.265568] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 85f5f90a-9591-4393-9bef-2ad7e6a1a82f/85f5f90a-9591-4393-9bef-2ad7e6a1a82f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 814.266488] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-27205592-0920-4076-9146-0ad4fa331e1a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.282801] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 814.282801] env[62000]: value = "task-882266" [ 814.282801] env[62000]: _type = "Task" [ 814.282801] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.301218] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882266, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.329588] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882264, 'name': CreateVM_Task, 'duration_secs': 0.983453} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.331264] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 814.334120] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 814.334120] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 814.334120] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 814.334430] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 814.335218] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 814.335447] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 814.335849] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 814.335849] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 814.336013] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 814.336390] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 814.336449] env[62000]: DEBUG nova.virt.hardware [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 814.337269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.337907] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.337907] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 814.339937] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6be8dbf-d3ce-48a2-9a35-41e6666efd6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.343856] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0fd5e22-1eeb-4650-ac52-e2960a593574 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.347058] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a43d3fa-0157-4a19-a379-55c76fc11384 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.351649] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 814.353124] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-581b7591-2053-462c-873a-2dc275e2b69f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.361305] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc943a67-2eb6-4fe8-8ef4-f515df32e8da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.369901] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 814.369901] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5235547a-632a-b2ba-6caf-6a8c25727c39" [ 814.369901] env[62000]: _type = "Task" [ 814.369901] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.370329] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 814.370329] env[62000]: ERROR oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk due to incomplete transfer. [ 814.371807] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16439fcc-cfa0-46ce-8945-cf0b6f60f286 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.377392] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9ac0d826-332e-4a8f-b4d2-96afee91d992 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.398169] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5235547a-632a-b2ba-6caf-6a8c25727c39, 'name': SearchDatastore_Task, 'duration_secs': 0.012789} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.425544] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.425920] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 814.426196] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.426286] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.426412] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 814.426717] env[62000]: DEBUG oslo_vmware.rw_handles [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/520be7ad-811c-dab1-94ee-38263f10b767/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 814.426886] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Uploaded image e83c8ab9-d1dc-497c-bb7e-7cd42feab5a7 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 814.429487] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 814.430026] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b9fb8ee-fa7d-4ba8-9c92-84d4d11406b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.432335] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fb2c135b-f8b5-46a9-b09f-1329c610013a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.434229] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd97ac7b-1389-4764-99f1-a91ec3d8b795 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.443823] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c09530d-d947-4fa2-bee0-f48fe0a9cf30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.450140] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 814.450140] env[62000]: value = "task-882267" [ 814.450140] env[62000]: _type = "Task" [ 814.450140] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.450418] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 814.450586] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 814.451790] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-846bf74c-e1dd-4b60-986f-1bb89c0738c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.465521] env[62000]: DEBUG nova.compute.provider_tree [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.474256] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 814.474256] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528dfd8c-cbe0-d5fb-b3e0-d00da387860b" [ 814.474256] env[62000]: _type = "Task" [ 814.474256] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.474538] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882267, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.488197] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528dfd8c-cbe0-d5fb-b3e0-d00da387860b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.489608] env[62000]: INFO nova.virt.block_device [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Booting with volume 2b463ae3-a1c9-43a7-a945-3da06e9ebae1 at /dev/sda [ 814.549628] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2028aa09-031a-4ecf-8118-284e071c328f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.561611] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1869b087-49b1-430a-9e00-381e2d56dc35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.601104] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5414df86-a1ad-4456-88d4-a46077d133b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.611098] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1759f9-f45d-4df5-bd61-d219375e1129 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.623447] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 814.651404] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adb8b9c-2e63-47cc-b6a6-6b7d99bc3b47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.659987] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f6bddc-0fc6-487f-945d-cab30053dc34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.679298] env[62000]: DEBUG nova.virt.block_device [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating existing volume attachment record: 06f39c11-536c-4297-99b1-9ea9628970c9 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 814.686295] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d1d987c4-1180-4de1-9818-b3e95849762c tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.559s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.796955] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882266, 'name': Rename_Task, 'duration_secs': 0.193487} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.798173] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 814.798605] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-da9387a1-283c-4afe-be42-6c88b57c276d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.802673] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Successfully created port: 3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 814.809710] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 814.809710] env[62000]: value = "task-882268" [ 814.809710] env[62000]: _type = "Task" [ 814.809710] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 814.819693] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882268, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.963619] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882267, 'name': Destroy_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 814.968695] env[62000]: DEBUG nova.scheduler.client.report [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.988450] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528dfd8c-cbe0-d5fb-b3e0-d00da387860b, 'name': SearchDatastore_Task, 'duration_secs': 0.014227} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 814.988450] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2bf82a10-1076-466b-91a0-c5c26a654d18 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.993157] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-515167d1-0a88-4083-b014-249c351061a8" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.993157] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-515167d1-0a88-4083-b014-249c351061a8" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.993408] env[62000]: DEBUG nova.objects.instance [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'flavor' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 814.997415] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 814.997415] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ba01f8-ab0c-8c96-fa35-c2a5e2574aeb" [ 814.997415] env[62000]: _type = "Task" [ 814.997415] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.009155] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ba01f8-ab0c-8c96-fa35-c2a5e2574aeb, 'name': SearchDatastore_Task, 'duration_secs': 0.010225} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.009624] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.010013] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 815.010432] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38bbd2d7-5e11-4226-8f34-c2a4b6a2e194 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.019712] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 815.019712] env[62000]: value = "task-882269" [ 815.019712] env[62000]: _type = "Task" [ 815.019712] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.029141] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882269, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.213802] env[62000]: DEBUG nova.network.neutron [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.326964] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.465681] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882267, 'name': Destroy_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.474159] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.511s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.474340] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 815.476983] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 2.989s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.477181] env[62000]: DEBUG nova.objects.instance [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 815.538897] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882269, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.720024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.720024] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Instance network_info: |[{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 815.720024] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:07:7f:55', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '849930d2-4200-4ec9-ac15-51c243a118bf', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 815.727731] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Creating folder: Project (6742b3d48eb945fe8a11d0ee651abd1a). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 815.728305] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b1449e9c-e472-455e-87d5-953a1faf91b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.748309] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Created folder: Project (6742b3d48eb945fe8a11d0ee651abd1a) in parent group-v201431. [ 815.748959] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Creating folder: Instances. Parent ref: group-v201553. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 815.749679] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-393da3e3-7a9d-4208-aab6-f20ed699a434 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.769107] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Created folder: Instances in parent group-v201553. [ 815.772631] env[62000]: DEBUG oslo.service.loopingcall [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 815.772631] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 815.772631] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d68a9dca-a6d1-47c6-ab99-71d6a44f4344 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.798102] env[62000]: DEBUG nova.objects.instance [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'pci_requests' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 815.810931] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 815.810931] env[62000]: value = "task-882272" [ 815.810931] env[62000]: _type = "Task" [ 815.810931] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.826139] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882272, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.831619] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882268, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.974621] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882267, 'name': Destroy_Task, 'duration_secs': 1.378163} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.974838] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroyed the VM [ 815.975077] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 815.975393] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b9a0fd60-8284-4ed4-987b-d43e465fb2b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.985407] env[62000]: DEBUG nova.compute.utils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 815.995201] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 815.995574] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 815.999372] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 815.999372] env[62000]: value = "task-882273" [ 815.999372] env[62000]: _type = "Task" [ 815.999372] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.021734] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882273, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.036684] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882269, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.077404] env[62000]: DEBUG nova.policy [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.080222] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.080583] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.305288] env[62000]: DEBUG nova.objects.base [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Object Instance<71192360-6ee6-4876-bf37-da987a09cbb2> lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 816.305686] env[62000]: DEBUG nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 816.329742] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882272, 'name': CreateVM_Task, 'duration_secs': 0.518907} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.334993] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 816.336027] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882268, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.337250] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.337621] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.338278] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 816.338945] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-76d289ff-43aa-4100-9b1c-d874f2d1507d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.350294] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 816.350294] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52121f3c-81d6-3b33-503b-f430e42e6b93" [ 816.350294] env[62000]: _type = "Task" [ 816.350294] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.364066] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52121f3c-81d6-3b33-503b-f430e42e6b93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.464372] env[62000]: DEBUG nova.policy [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 816.495871] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 816.507029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-caa91042-b80a-4271-a2ae-dac5b253e9b0 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.028s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.507029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.567s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.507029] env[62000]: DEBUG nova.objects.instance [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'resources' on Instance uuid ae343199-1006-473a-a47f-7983835bd60a {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 816.519820] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882273, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.536438] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882269, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.336058} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.538039] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 816.538357] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 816.538989] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d061583b-7b26-4f16-b5b0-a47812261019 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.550382] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 816.550382] env[62000]: value = "task-882274" [ 816.550382] env[62000]: _type = "Task" [ 816.550382] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.562020] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882274, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.584118] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 816.601173] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Successfully created port: 768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 816.807901] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.808181] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.826511] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 816.826953] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 816.827208] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 816.827383] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 816.827571] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 816.827734] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 816.827925] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 816.830813] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 816.831246] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 816.831447] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 816.831644] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 816.831823] env[62000]: DEBUG nova.virt.hardware [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 816.844775] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a683dc7c-d251-422c-8fea-d85d80ec62a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.849524] env[62000]: DEBUG oslo_vmware.api [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882268, 'name': PowerOnVM_Task, 'duration_secs': 1.831733} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.850188] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 816.850403] env[62000]: INFO nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Took 12.24 seconds to spawn the instance on the hypervisor. [ 816.850581] env[62000]: DEBUG nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 816.852442] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be846a1f-b558-4470-9d9a-d98048ea49d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.865342] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae347a8-72dc-4306-b3e6-7a918f11fbd2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.871706] env[62000]: DEBUG nova.compute.manager [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Received event network-changed-d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 816.871905] env[62000]: DEBUG nova.compute.manager [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Refreshing instance network info cache due to event network-changed-d8f74268-0a99-4e0d-8042-52662b7e2c21. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 816.872191] env[62000]: DEBUG oslo_concurrency.lockutils [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] Acquiring lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.872357] env[62000]: DEBUG oslo_concurrency.lockutils [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] Acquired lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.872523] env[62000]: DEBUG nova.network.neutron [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Refreshing network info cache for port d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 816.885516] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52121f3c-81d6-3b33-503b-f430e42e6b93, 'name': SearchDatastore_Task, 'duration_secs': 0.012058} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.886129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.886283] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 816.886533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.886650] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.890272] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 816.902288] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e1618abe-0fcb-41b0-967d-1358ef3b97a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.917995] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 816.918140] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 816.919076] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e846aca-80a5-49c6-9271-648e21d33601 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.929024] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 816.929024] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5243dcd6-9f64-91a8-79c0-17affe251dce" [ 816.929024] env[62000]: _type = "Task" [ 816.929024] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 816.938877] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5243dcd6-9f64-91a8-79c0-17affe251dce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.022194] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882273, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.065961] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882274, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.113789} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.066551] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 817.067479] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e9167e1-2d60-4310-93f1-144f3034d113 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.093510] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 817.098910] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9eb4aebb-56e4-4d5d-b63e-b183b13389f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.122931] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 817.122931] env[62000]: value = "task-882275" [ 817.122931] env[62000]: _type = "Task" [ 817.122931] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.132765] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882275, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.133959] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.314947] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.333795] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c000ef-5b0a-417e-9793-0328cbbab0ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.344684] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8c7fd5-c431-4d69-a09d-65619922ca69 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.389722] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Successfully updated port: 3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 817.395994] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5017d37-bb3d-4d71-a063-8c57b38af268 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.400549] env[62000]: INFO nova.compute.manager [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Took 36.11 seconds to build instance. [ 817.408828] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bf5f56-496f-415f-8096-70901278f416 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.428023] env[62000]: DEBUG nova.compute.provider_tree [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.441258] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5243dcd6-9f64-91a8-79c0-17affe251dce, 'name': SearchDatastore_Task, 'duration_secs': 0.023116} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.443446] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 817.447242] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fe4397-ed74-40f0-8978-ed7821b62c41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.451398] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c8e9e869-6fe8-4a3f-9ca4-e47a531881b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.459616] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 817.459616] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977852-71bb-15ab-fdc3-33cdcaf971e5" [ 817.459616] env[62000]: _type = "Task" [ 817.459616] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.461370] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 817.461549] env[62000]: ERROR oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk due to incomplete transfer. [ 817.464937] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9a90b3b0-e75c-45c7-a733-3bcbe386e1fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.473603] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977852-71bb-15ab-fdc3-33cdcaf971e5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.474996] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d4c933-4d7a-11ac-cca0-1b5c5ff25576/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 817.475207] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Uploaded image 53a44733-1c2c-4740-b053-0d40c24bf75b to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 817.476811] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 817.477422] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-04373c1b-65a5-44ef-b3c4-ffb67ef5bd21 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.484194] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 817.484194] env[62000]: value = "task-882276" [ 817.484194] env[62000]: _type = "Task" [ 817.484194] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.492478] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882276, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.508253] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 817.525804] env[62000]: DEBUG oslo_vmware.api [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882273, 'name': RemoveSnapshot_Task, 'duration_secs': 1.138021} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.526120] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 817.527545] env[62000]: INFO nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 16.89 seconds to snapshot the instance on the hypervisor. [ 817.536726] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.536967] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.537146] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.537342] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.537482] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.537630] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.537837] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.538010] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.538187] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.538341] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.538548] env[62000]: DEBUG nova.virt.hardware [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.539545] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d24269a4-e85c-466e-8cf0-5f62c75f825b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.547661] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2971c56-d21a-4520-af56-a00446158917 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.634224] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882275, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.699165] env[62000]: DEBUG nova.network.neutron [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Updated VIF entry in instance network info cache for port d8f74268-0a99-4e0d-8042-52662b7e2c21. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 817.699165] env[62000]: DEBUG nova.network.neutron [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Updating instance_info_cache with network_info: [{"id": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "address": "fa:16:3e:b6:d0:9f", "network": {"id": "2d839574-e381-4fdf-970c-69737dbc35dc", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1373885955-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "817c2729a04843518a4333c6203a29b7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f9be6786-e9a7-4138-b7b5-b7696f6cb1e1", "external-id": "nsx-vlan-transportzone-626", "segmentation_id": 626, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8f74268-0a", "ovs_interfaceid": "d8f74268-0a99-4e0d-8042-52662b7e2c21", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.845743] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.900655] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.900791] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquired lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.900957] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.907185] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4179977f-2f0c-4ab6-a14b-9765981778a0 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.620s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.933964] env[62000]: DEBUG nova.scheduler.client.report [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.974100] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977852-71bb-15ab-fdc3-33cdcaf971e5, 'name': SearchDatastore_Task, 'duration_secs': 0.038879} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.974100] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.974100] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 66edb906-f936-4f57-833d-224f36af109e/66edb906-f936-4f57-833d-224f36af109e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 817.974100] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-64dbd55b-d547-4d58-a4f4-d9bbf8b0d3a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.984918] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 817.984918] env[62000]: value = "task-882277" [ 817.984918] env[62000]: _type = "Task" [ 817.984918] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.001508] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882276, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.005845] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882277, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.093236] env[62000]: DEBUG nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Found 3 images (rotation: 2) {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 818.093498] env[62000]: DEBUG nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Rotating out 1 backups {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 818.093766] env[62000]: DEBUG nova.compute.manager [None req-605e50d3-d629-473e-972c-f53fc3bdba1b tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleting image 150ffec2-eb6a-4b9b-baf4-6005539ac9fa {{(pid=62000) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 818.137742] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882275, 'name': ReconfigVM_Task, 'duration_secs': 1.013588} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.138764] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Reconfigured VM instance instance-0000003f to attach disk [datastore1] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 818.139737] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-72add919-762d-4c0a-8806-fe7cab55ad62 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.149691] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 818.149691] env[62000]: value = "task-882278" [ 818.149691] env[62000]: _type = "Task" [ 818.149691] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.159452] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882278, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.203784] env[62000]: DEBUG oslo_concurrency.lockutils [req-961076af-c148-49f5-a32e-2a79894d5974 req-bb0efe83-5757-4473-8af4-580725505e09 service nova] Releasing lock "refresh_cache-85f5f90a-9591-4393-9bef-2ad7e6a1a82f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.282665] env[62000]: DEBUG nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Successfully updated port: 515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.383591] env[62000]: DEBUG nova.compute.manager [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Received event network-vif-plugged-0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.383706] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.383911] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.384974] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.385296] env[62000]: DEBUG nova.compute.manager [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] No waiting events found dispatching network-vif-plugged-0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 818.385598] env[62000]: WARNING nova.compute.manager [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Received unexpected event network-vif-plugged-0cb92206-16ee-49e1-8e70-fa3f39d678b4 for instance with vm_state building and task_state spawning. [ 818.385973] env[62000]: DEBUG nova.compute.manager [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Received event network-changed-0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 818.386224] env[62000]: DEBUG nova.compute.manager [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Refreshing instance network info cache due to event network-changed-0cb92206-16ee-49e1-8e70-fa3f39d678b4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 818.386509] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Acquiring lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.386705] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Acquired lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.386963] env[62000]: DEBUG nova.network.neutron [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Refreshing network info cache for port 0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 818.439657] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.933s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.442139] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 4.283s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.452322] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.469875] env[62000]: INFO nova.scheduler.client.report [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocations for instance ae343199-1006-473a-a47f-7983835bd60a [ 818.501673] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882276, 'name': Destroy_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.505638] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882277, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.624271] env[62000]: DEBUG nova.network.neutron [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating instance_info_cache with network_info: [{"id": "3c405e0f-e997-4845-a292-10cad3943d50", "address": "fa:16:3e:3f:0b:64", "network": {"id": "14ad3cf7-b997-4f08-8510-945d34e17eac", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-315454476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4b4854b30648d8ac861f08fbdb6da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c405e0f-e9", "ovs_interfaceid": "3c405e0f-e997-4845-a292-10cad3943d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.660864] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882278, 'name': Rename_Task, 'duration_secs': 0.197378} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.661656] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 818.661984] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee37e811-e51b-4c77-9114-93c46781f559 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.670400] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 818.670400] env[62000]: value = "task-882279" [ 818.670400] env[62000]: _type = "Task" [ 818.670400] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.682301] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882279, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.706511] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Successfully updated port: 768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 818.785155] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.785412] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.785568] env[62000]: DEBUG nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 818.945227] env[62000]: DEBUG nova.objects.instance [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lazy-loading 'migration_context' on Instance uuid f3a6527c-777e-4a07-9482-598de15d4eb3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 818.985253] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e5d881d-f772-45f6-8386-de9c123b697e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "ae343199-1006-473a-a47f-7983835bd60a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.411s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.999333] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882276, 'name': Destroy_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.006314] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882277, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.734085} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.009820] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 66edb906-f936-4f57-833d-224f36af109e/66edb906-f936-4f57-833d-224f36af109e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 819.009820] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.010054] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4f68211d-8802-4de5-8523-91920546f909 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.022021] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 819.022021] env[62000]: value = "task-882280" [ 819.022021] env[62000]: _type = "Task" [ 819.022021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.031161] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882280, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.037928] env[62000]: DEBUG nova.objects.instance [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lazy-loading 'flavor' on Instance uuid ea8ae421-1d7f-4814-bc0f-90a3316ad028 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 819.127210] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Releasing lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.127701] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance network_info: |[{"id": "3c405e0f-e997-4845-a292-10cad3943d50", "address": "fa:16:3e:3f:0b:64", "network": {"id": "14ad3cf7-b997-4f08-8510-945d34e17eac", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-315454476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4b4854b30648d8ac861f08fbdb6da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c405e0f-e9", "ovs_interfaceid": "3c405e0f-e997-4845-a292-10cad3943d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 819.128546] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3f:0b:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '97b68ed7-8461-4345-b064-96a1dde53a86', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3c405e0f-e997-4845-a292-10cad3943d50', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 819.138694] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Creating folder: Project (7d4b4854b30648d8ac861f08fbdb6da9). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.141387] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69e981ea-79b5-4f2f-a99d-acf4bb133f82 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.162282] env[62000]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 819.162455] env[62000]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62000) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 819.162876] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Folder already exists: Project (7d4b4854b30648d8ac861f08fbdb6da9). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 819.163116] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Creating folder: Instances. Parent ref: group-v201532. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 819.163395] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9311bf64-dcf6-44bc-9e4c-080ebb673173 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.178215] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Created folder: Instances in parent group-v201532. [ 819.178504] env[62000]: DEBUG oslo.service.loopingcall [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 819.181942] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 819.182286] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882279, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.182509] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8dbb3e65-72be-4f1c-8abf-e7575a28a186 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.201229] env[62000]: DEBUG nova.network.neutron [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Updated VIF entry in instance network info cache for port 0cb92206-16ee-49e1-8e70-fa3f39d678b4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 819.202025] env[62000]: DEBUG nova.network.neutron [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Updating instance_info_cache with network_info: [{"id": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "address": "fa:16:3e:03:c1:8c", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cb92206-16", "ovs_interfaceid": "0cb92206-16ee-49e1-8e70-fa3f39d678b4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.204968] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.205056] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.206132] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 819.213167] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 819.213167] env[62000]: value = "task-882283" [ 819.213167] env[62000]: _type = "Task" [ 819.213167] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.224252] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882283, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.384488] env[62000]: WARNING nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] e21c3795-1c6b-42ef-af81-e113912fa80d already exists in list: networks containing: ['e21c3795-1c6b-42ef-af81-e113912fa80d']. ignoring it [ 819.384742] env[62000]: WARNING nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] e21c3795-1c6b-42ef-af81-e113912fa80d already exists in list: networks containing: ['e21c3795-1c6b-42ef-af81-e113912fa80d']. ignoring it [ 819.500043] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882276, 'name': Destroy_Task, 'duration_secs': 1.808107} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.500408] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Destroyed the VM [ 819.500726] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 819.501077] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-841e15fa-4b71-4feb-bfe1-fbf18c785533 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.522205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.522205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.523846] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 819.523846] env[62000]: value = "task-882284" [ 819.523846] env[62000]: _type = "Task" [ 819.523846] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.548978] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882280, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.104468} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.554672] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.554872] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.557469] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 819.557757] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882284, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.558803] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142182ac-a275-4513-861c-ce9a0cbf0409 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.590321] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] 66edb906-f936-4f57-833d-224f36af109e/66edb906-f936-4f57-833d-224f36af109e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 819.593855] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7dd971b9-9668-44ed-b545-a05e9c9c4bde {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.618317] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 819.618317] env[62000]: value = "task-882285" [ 819.618317] env[62000]: _type = "Task" [ 819.618317] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.630757] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882285, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.686126] env[62000]: DEBUG oslo_vmware.api [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882279, 'name': PowerOnVM_Task, 'duration_secs': 0.854269} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.686367] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 819.686684] env[62000]: INFO nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Took 12.65 seconds to spawn the instance on the hypervisor. [ 819.686944] env[62000]: DEBUG nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 819.687911] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e363f5-9bb4-4331-b87d-8bdb77034354 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.705134] env[62000]: DEBUG oslo_concurrency.lockutils [req-08e6a50c-f9a1-4593-a5f1-d830d224187f req-4a9f39ca-c68a-4355-b0c2-d0cd194f82ae service nova] Releasing lock "refresh_cache-e2e3270d-086b-4441-a3d6-49b05a60b51f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.726056] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882283, 'name': CreateVM_Task, 'duration_secs': 0.477537} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.726245] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 819.727056] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '06f39c11-536c-4297-99b1-9ea9628970c9', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201541', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'name': 'volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2f35dab3-6c32-4a35-91ca-1d6859e27f2c', 'attached_at': '', 'detached_at': '', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'serial': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1'}, 'mount_device': '/dev/sda', 'device_type': None, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62000) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 819.727301] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Root volume attach. Driver type: vmdk {{(pid=62000) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 819.728182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74068591-4738-44da-a0fb-07ffbb9955f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.741459] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62c8e15-2b47-4e98-aaa7-2d97ce93259d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.751626] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba99f1d-264e-4540-8964-9c186ebdaa0f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.760200] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-58d9070c-77db-4e3d-8a6f-9b19cae9bc52 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.776868] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.782888] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 819.782888] env[62000]: value = "task-882286" [ 819.782888] env[62000]: _type = "Task" [ 819.782888] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.793325] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882286, 'name': RelocateVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.827325] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d761f7e9-83b2-4672-aef9-4c456d8662e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.839939] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00f110a4-6fd2-4870-8b6c-4b809b0a44e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.874753] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223af20b-9503-4f50-af52-a4b082a1388e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.883851] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc4f1e2-f90a-4266-b6e2-9d354a3c8987 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.906276] env[62000]: DEBUG nova.compute.provider_tree [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.025571] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.041876] env[62000]: DEBUG oslo_vmware.api [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882284, 'name': RemoveSnapshot_Task, 'duration_secs': 0.504826} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.045167] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 820.045666] env[62000]: INFO nova.compute.manager [None req-d8c8c534-3c75-4fca-bdcc-b988462ec93c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Took 17.60 seconds to snapshot the instance on the hypervisor. [ 820.132159] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882285, 'name': ReconfigVM_Task, 'duration_secs': 0.467887} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.132159] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfigured VM instance instance-00000040 to attach disk [datastore2] 66edb906-f936-4f57-833d-224f36af109e/66edb906-f936-4f57-833d-224f36af109e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.133057] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d3f6a5f7-1bcb-4b4c-a966-0ca8bbb23920 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.145111] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 820.145111] env[62000]: value = "task-882287" [ 820.145111] env[62000]: _type = "Task" [ 820.145111] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.157029] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882287, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.211732] env[62000]: INFO nova.compute.manager [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Took 29.19 seconds to build instance. [ 820.294864] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882286, 'name': RelocateVM_Task} progress is 20%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.412194] env[62000]: DEBUG nova.scheduler.client.report [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 820.453404] env[62000]: DEBUG nova.network.neutron [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Updating instance_info_cache with network_info: [{"id": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "address": "fa:16:3e:52:df:58", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap768a74c4-85", "ovs_interfaceid": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.494161] env[62000]: DEBUG nova.network.neutron [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.563140] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.660188] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882287, 'name': Rename_Task, 'duration_secs': 0.328027} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.662048] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 820.662048] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d40ae353-bb2d-4d9c-bbcc-c7a396a542b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.672368] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 820.672368] env[62000]: value = "task-882288" [ 820.672368] env[62000]: _type = "Task" [ 820.672368] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.681948] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882288, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.714103] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fd652c2f-408a-4f6b-b267-179c9cb48250 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.704s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.786258] env[62000]: DEBUG nova.network.neutron [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "656e8902-9da0-40c9-889a-8d1aab0f9266", "address": "fa:16:3e:24:53:f9", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap656e8902-9d", "ovs_interfaceid": "656e8902-9da0-40c9-889a-8d1aab0f9266", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "515167d1-0a88-4083-b014-249c351061a8", "address": "fa:16:3e:16:52:5d", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515167d1-0a", "ovs_interfaceid": "515167d1-0a88-4083-b014-249c351061a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.797532] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882286, 'name': RelocateVM_Task, 'duration_secs': 0.793537} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.798066] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 820.798390] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201541', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'name': 'volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2f35dab3-6c32-4a35-91ca-1d6859e27f2c', 'attached_at': '', 'detached_at': '', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'serial': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 820.799344] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e52dda-0b19-4b34-8382-ed20ad28e64a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.821179] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e0ef00-1b31-46cf-9628-d1910354f20d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.847817] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1/volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.848303] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c5b947d-c3a4-4718-844c-772545ebf10c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.872852] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 820.872852] env[62000]: value = "task-882289" [ 820.872852] env[62000]: _type = "Task" [ 820.872852] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.882589] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882289, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.930510] env[62000]: DEBUG nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: ae343199-1006-473a-a47f-7983835bd60a] Received event network-vif-deleted-00896a83-2b1c-4c70-86a9-243a6ccbc36f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.930714] env[62000]: DEBUG nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-plugged-515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.931992] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.931992] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.931992] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.931992] env[62000]: DEBUG nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] No waiting events found dispatching network-vif-plugged-515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 820.931992] env[62000]: WARNING nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received unexpected event network-vif-plugged-515167d1-0a88-4083-b014-249c351061a8 for instance with vm_state active and task_state None. [ 820.931992] env[62000]: DEBUG nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-changed-515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 820.931992] env[62000]: DEBUG nova.compute.manager [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing instance network info cache due to event network-changed-515167d1-0a88-4083-b014-249c351061a8. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 820.932281] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.956119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.956457] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Instance network_info: |[{"id": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "address": "fa:16:3e:52:df:58", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap768a74c4-85", "ovs_interfaceid": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 820.957018] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:df:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '768a74c4-85dd-4101-bd5a-11290b69e3a8', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 820.965603] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating folder: Project (43e5c47870584d05abaf9de72d45cce2). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.966453] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0638b228-d246-45dd-9b90-031f14c8d1ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.982673] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created folder: Project (43e5c47870584d05abaf9de72d45cce2) in parent group-v201431. [ 820.982673] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating folder: Instances. Parent ref: group-v201558. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 820.982673] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8cdcea9a-0890-427f-a6d6-76b385c20a35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.994718] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created folder: Instances in parent group-v201558. [ 820.995221] env[62000]: DEBUG oslo.service.loopingcall [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.995328] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 820.995579] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1a4c361e-f92d-464b-9a0c-3d6389b61248 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.020670] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 821.020670] env[62000]: value = "task-882292" [ 821.020670] env[62000]: _type = "Task" [ 821.020670] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.030530] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882292, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.182698] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882288, 'name': PowerOnVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.293782] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.294730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.294730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.295173] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.295337] env[62000]: DEBUG nova.network.neutron [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Refreshing network info cache for port 515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 821.301021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39f9a155-e3db-404a-84b4-244036b3b18c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.322279] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.322955] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.322955] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.323159] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.323248] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.323450] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.324035] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.324035] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.324175] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.324356] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.324649] env[62000]: DEBUG nova.virt.hardware [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.331837] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfiguring VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 821.333228] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-449dfdc7-11fd-49df-849e-597a5e78e347 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.368364] env[62000]: DEBUG oslo_vmware.api [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 821.368364] env[62000]: value = "task-882293" [ 821.368364] env[62000]: _type = "Task" [ 821.368364] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.382214] env[62000]: DEBUG oslo_vmware.api [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882293, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.388884] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882289, 'name': ReconfigVM_Task, 'duration_secs': 0.373487} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.389212] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Reconfigured VM instance instance-00000041 to attach disk [datastore2] volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1/volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 821.394549] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff027f20-d941-4768-a292-cfc33663cfcf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.415435] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 821.415435] env[62000]: value = "task-882294" [ 821.415435] env[62000]: _type = "Task" [ 821.415435] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.429226] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882294, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.435555] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.993s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.441232] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.307s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.442769] env[62000]: INFO nova.compute.claims [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.536055] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882292, 'name': CreateVM_Task, 'duration_secs': 0.433483} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.536417] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 821.537330] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.537748] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.538232] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 821.538636] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-528bb650-85c3-41f6-8cb8-fb0b53e33c19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.547074] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 821.547074] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e705a2-b1b7-18fb-1c84-206895d7ba17" [ 821.547074] env[62000]: _type = "Task" [ 821.547074] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.559068] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e705a2-b1b7-18fb-1c84-206895d7ba17, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.592191] env[62000]: DEBUG nova.network.neutron [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.682893] env[62000]: DEBUG oslo_vmware.api [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882288, 'name': PowerOnVM_Task, 'duration_secs': 0.66079} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.683221] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 821.683423] env[62000]: INFO nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Took 10.13 seconds to spawn the instance on the hypervisor. [ 821.683606] env[62000]: DEBUG nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.684440] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e70e8a9-0df6-4fdd-a38d-4e2c3c0aaeb5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.791129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.792052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.792052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "153b86d9-dc2e-463d-87fd-155ec23e2abb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.792052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.792052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.794289] env[62000]: INFO nova.compute.manager [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Terminating instance [ 821.796562] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "refresh_cache-153b86d9-dc2e-463d-87fd-155ec23e2abb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.796721] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquired lock "refresh_cache-153b86d9-dc2e-463d-87fd-155ec23e2abb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.797256] env[62000]: DEBUG nova.network.neutron [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 821.881927] env[62000]: DEBUG oslo_vmware.api [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882293, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.925977] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882294, 'name': ReconfigVM_Task, 'duration_secs': 0.176596} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.926324] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201541', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'name': 'volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2f35dab3-6c32-4a35-91ca-1d6859e27f2c', 'attached_at': '', 'detached_at': '', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'serial': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 821.926909] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5663662e-03fb-41e7-a32b-bfaa25e827dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.935814] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 821.935814] env[62000]: value = "task-882295" [ 821.935814] env[62000]: _type = "Task" [ 821.935814] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.945233] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882295, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.990982] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Received event network-vif-plugged-849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.992050] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.992877] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "66edb906-f936-4f57-833d-224f36af109e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.993102] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "66edb906-f936-4f57-833d-224f36af109e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.993344] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] No waiting events found dispatching network-vif-plugged-849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 821.993558] env[62000]: WARNING nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Received unexpected event network-vif-plugged-849930d2-4200-4ec9-ac15-51c243a118bf for instance with vm_state active and task_state None. [ 821.993947] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Received event network-changed-849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 821.994173] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Refreshing instance network info cache due to event network-changed-849930d2-4200-4ec9-ac15-51c243a118bf. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 821.994372] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.994830] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquired lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.994830] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Refreshing network info cache for port 849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 822.063813] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e705a2-b1b7-18fb-1c84-206895d7ba17, 'name': SearchDatastore_Task, 'duration_secs': 0.013914} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.064604] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.065356] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 822.065503] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.065655] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.065858] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 822.066581] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1bf353d0-6c3b-462f-900f-434e67f2ef5d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.080349] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 822.080679] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 822.081694] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ea7cdee-0947-4040-ac8e-96490c312d25 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.092167] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 822.092167] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5245bcb3-0588-4fb6-2391-973e7cb59d2f" [ 822.092167] env[62000]: _type = "Task" [ 822.092167] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.096280] env[62000]: DEBUG oslo_concurrency.lockutils [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.096537] env[62000]: DEBUG nova.compute.manager [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Inject network info {{(pid=62000) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 822.096794] env[62000]: DEBUG nova.compute.manager [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] network_info to inject: |[{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 822.102742] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfiguring VM instance to set the machine id {{(pid=62000) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 822.103581] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ddf83f94-812f-4007-97f8-11fd16a18171 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.118863] env[62000]: DEBUG nova.network.neutron [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updated VIF entry in instance network info cache for port 515167d1-0a88-4083-b014-249c351061a8. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.119420] env[62000]: DEBUG nova.network.neutron [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "656e8902-9da0-40c9-889a-8d1aab0f9266", "address": "fa:16:3e:24:53:f9", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap656e8902-9d", "ovs_interfaceid": "656e8902-9da0-40c9-889a-8d1aab0f9266", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "515167d1-0a88-4083-b014-249c351061a8", "address": "fa:16:3e:16:52:5d", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515167d1-0a", "ovs_interfaceid": "515167d1-0a88-4083-b014-249c351061a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.127998] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5245bcb3-0588-4fb6-2391-973e7cb59d2f, 'name': SearchDatastore_Task, 'duration_secs': 0.012303} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.130428] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8a823cd5-d6fd-4cf6-adc7-5b795d5938f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.133802] env[62000]: DEBUG oslo_vmware.api [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 822.133802] env[62000]: value = "task-882296" [ 822.133802] env[62000]: _type = "Task" [ 822.133802] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.144987] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 822.144987] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c5fedd-9b32-8f55-bcec-ddf842fb2cff" [ 822.144987] env[62000]: _type = "Task" [ 822.144987] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.154107] env[62000]: DEBUG oslo_vmware.api [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882296, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.161526] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c5fedd-9b32-8f55-bcec-ddf842fb2cff, 'name': SearchDatastore_Task, 'duration_secs': 0.013812} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.161798] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.162129] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 48d29490-c174-4702-a449-3bf10bc9207c/48d29490-c174-4702-a449-3bf10bc9207c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 822.162412] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ebd8c8d4-2f3a-4296-b317-a40a720d318d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.171452] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 822.171452] env[62000]: value = "task-882297" [ 822.171452] env[62000]: _type = "Task" [ 822.171452] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.181332] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882297, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.202621] env[62000]: INFO nova.compute.manager [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Took 26.78 seconds to build instance. [ 822.321156] env[62000]: DEBUG nova.network.neutron [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.386026] env[62000]: DEBUG oslo_vmware.api [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882293, 'name': ReconfigVM_Task, 'duration_secs': 0.974964} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.386026] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.386234] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfigured VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 822.408202] env[62000]: DEBUG nova.network.neutron [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.449618] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882295, 'name': Rename_Task, 'duration_secs': 0.201868} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.450052] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 822.450437] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cacd1758-2b1d-4151-ac37-5b4da4124e4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.465948] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 822.465948] env[62000]: value = "task-882298" [ 822.465948] env[62000]: _type = "Task" [ 822.465948] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.486625] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882298, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.624228] env[62000]: DEBUG oslo_concurrency.lockutils [req-61c31c13-115c-4ffd-8b94-356ec32ca188 req-777f0794-6fd4-4f6a-a709-b6e17e18d174 service nova] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.651934] env[62000]: DEBUG oslo_vmware.api [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882296, 'name': ReconfigVM_Task, 'duration_secs': 0.17882} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 822.651934] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-4fa0d1ca-6191-4665-95cc-99a2534a14fc tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfigured VM instance to set the machine id {{(pid=62000) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 822.688610] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882297, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.705626] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e9ea633-9ece-454b-983e-e9a2208b3e22 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.294s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.801659] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d964f61c-6feb-48be-8691-b5001e244d4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.811996] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b91135-cf81-46bc-8387-979946b03f79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.852370] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65852902-99de-4b69-ad83-fd6648a52ca1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.860551] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0834f7cb-5a66-4cdb-bfe7-79970ed1989a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.866180] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updated VIF entry in instance network info cache for port 849930d2-4200-4ec9-ac15-51c243a118bf. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 822.866568] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.875833] env[62000]: DEBUG nova.compute.provider_tree [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 822.893930] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0e3ebde9-8e93-4a41-8253-98d0265bbe54 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-515167d1-0a88-4083-b014-249c351061a8" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.901s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.911252] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Releasing lock "refresh_cache-153b86d9-dc2e-463d-87fd-155ec23e2abb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.911703] env[62000]: DEBUG nova.compute.manager [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 822.911909] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 822.912985] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb57f132-83f0-49e6-95fb-1c880ba6f476 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.923531] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 822.923809] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4ffd35e7-b204-4db5-9f60-1e887890fd48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.931835] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 822.931835] env[62000]: value = "task-882299" [ 822.931835] env[62000]: _type = "Task" [ 822.931835] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 822.941706] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882299, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 822.990192] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882298, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.000181] env[62000]: INFO nova.compute.manager [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Swapping old allocation on dict_keys(['bcfccf20-49dd-4b91-819e-4373e67bf5ec']) held by migration 5e0eeb7d-c7f1-4c3a-b730-71573fb528fc for instance [ 823.032980] env[62000]: DEBUG nova.scheduler.client.report [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Overwriting current allocation {'allocations': {'bcfccf20-49dd-4b91-819e-4373e67bf5ec': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 78}}, 'project_id': '45a0491d5e13450bb11b99c4fba2807f', 'user_id': '147e732bf4e94249abef07fd2090ea80', 'consumer_generation': 1} on consumer f3a6527c-777e-4a07-9482-598de15d4eb3 {{(pid=62000) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 823.183554] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882297, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.652088} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.184186] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 48d29490-c174-4702-a449-3bf10bc9207c/48d29490-c174-4702-a449-3bf10bc9207c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 823.184661] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 823.185059] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-461b4d43-b3c3-472a-b7a8-90b48b48c86b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.197374] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 823.197374] env[62000]: value = "task-882300" [ 823.197374] env[62000]: _type = "Task" [ 823.197374] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.209490] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882300, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.212721] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.213012] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquired lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.213340] env[62000]: DEBUG nova.network.neutron [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 823.379796] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Releasing lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.380284] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Received event network-vif-plugged-3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.380499] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.380710] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.380880] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.381067] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] No waiting events found dispatching network-vif-plugged-3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 823.381248] env[62000]: WARNING nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Received unexpected event network-vif-plugged-3c405e0f-e997-4845-a292-10cad3943d50 for instance with vm_state building and task_state spawning. [ 823.381483] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Received event network-changed-3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 823.381662] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Refreshing instance network info cache due to event network-changed-3c405e0f-e997-4845-a292-10cad3943d50. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 823.381856] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.382090] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquired lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.382282] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Refreshing network info cache for port 3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 823.384627] env[62000]: DEBUG nova.scheduler.client.report [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.445772] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882299, 'name': PowerOffVM_Task, 'duration_secs': 0.135338} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.445772] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 823.445772] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 823.445772] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2ba181b-a6b8-424b-9c78-78129be4f3bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.482222] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 823.483492] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 823.483492] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Deleting the datastore file [datastore1] 153b86d9-dc2e-463d-87fd-155ec23e2abb {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 823.484182] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72ffca83-e855-42fe-b69a-8837e4c9d654 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.493666] env[62000]: DEBUG oslo_vmware.api [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882298, 'name': PowerOnVM_Task, 'duration_secs': 1.006565} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.495840] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 823.495840] env[62000]: INFO nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Took 6.67 seconds to spawn the instance on the hypervisor. [ 823.495840] env[62000]: DEBUG nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 823.495840] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c9e181b-cd70-4ed8-bd2b-47dfdf5b1ed2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.500054] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for the task: (returnval){ [ 823.500054] env[62000]: value = "task-882302" [ 823.500054] env[62000]: _type = "Task" [ 823.500054] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.515847] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882302, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.708235] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882300, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077998} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 823.708594] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 823.709403] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41959731-7abb-46f2-8793-aa4c6884d9cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.735851] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 48d29490-c174-4702-a449-3bf10bc9207c/48d29490-c174-4702-a449-3bf10bc9207c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 823.736238] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d97f8b9a-f83c-4b29-a1f1-54aa24fd2605 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.759113] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 823.759113] env[62000]: value = "task-882303" [ 823.759113] env[62000]: _type = "Task" [ 823.759113] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 823.768973] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882303, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 823.893689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.452s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.895213] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.898364] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.053s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.900106] env[62000]: INFO nova.compute.claims [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.017345] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.017944] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.028095] env[62000]: DEBUG oslo_vmware.api [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Task: {'id': task-882302, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.358154} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.028721] env[62000]: INFO nova.compute.manager [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Took 14.20 seconds to build instance. [ 824.030217] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 824.030467] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 824.030826] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 824.032886] env[62000]: INFO nova.compute.manager [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 824.032886] env[62000]: DEBUG oslo.service.loopingcall [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 824.037622] env[62000]: DEBUG nova.compute.manager [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 824.037622] env[62000]: DEBUG nova.network.neutron [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 824.074498] env[62000]: DEBUG nova.network.neutron [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 824.252395] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updated VIF entry in instance network info cache for port 3c405e0f-e997-4845-a292-10cad3943d50. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 824.252912] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating instance_info_cache with network_info: [{"id": "3c405e0f-e997-4845-a292-10cad3943d50", "address": "fa:16:3e:3f:0b:64", "network": {"id": "14ad3cf7-b997-4f08-8510-945d34e17eac", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-315454476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4b4854b30648d8ac861f08fbdb6da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c405e0f-e9", "ovs_interfaceid": "3c405e0f-e997-4845-a292-10cad3943d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.268097] env[62000]: DEBUG nova.network.neutron [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [{"id": "ddd21181-32c0-4898-bd09-7689d6976198", "address": "fa:16:3e:e8:24:07", "network": {"id": "7b65f2c7-e5ef-47cd-b442-80609ed2e052", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.103", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "29e4579cf2604191945dca831f024a21", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7c09cc47-a7d0-4816-bee4-69cc9f2e04b0", "external-id": "nsx-vlan-transportzone-687", "segmentation_id": 687, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapddd21181-32", "ovs_interfaceid": "ddd21181-32c0-4898-bd09-7689d6976198", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.273258] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882303, 'name': ReconfigVM_Task, 'duration_secs': 0.334961} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.274125] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 48d29490-c174-4702-a449-3bf10bc9207c/48d29490-c174-4702-a449-3bf10bc9207c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 824.274793] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c72adb3-c64d-451c-aa85-ae3691effbcb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.284777] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 824.284777] env[62000]: value = "task-882304" [ 824.284777] env[62000]: _type = "Task" [ 824.284777] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.300378] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882304, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.405566] env[62000]: DEBUG nova.compute.utils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.414138] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.414138] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 824.479896] env[62000]: DEBUG nova.policy [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4770851cc3b4ad8aaa6866a18ee1359', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41e10f7d4f964f2795dc629721802880', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.528704] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.539114] env[62000]: DEBUG oslo_concurrency.lockutils [None req-847ac5cb-424d-4c0c-99f0-e996b0be8bea tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.727s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.578328] env[62000]: DEBUG nova.network.neutron [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.678688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.678688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.757758] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Releasing lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.758289] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Received event network-vif-plugged-768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.758511] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "48d29490-c174-4702-a449-3bf10bc9207c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.758724] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "48d29490-c174-4702-a449-3bf10bc9207c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.758895] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Lock "48d29490-c174-4702-a449-3bf10bc9207c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.761089] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] No waiting events found dispatching network-vif-plugged-768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 824.761399] env[62000]: WARNING nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Received unexpected event network-vif-plugged-768a74c4-85dd-4101-bd5a-11290b69e3a8 for instance with vm_state building and task_state spawning. [ 824.761509] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Received event network-changed-768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.761675] env[62000]: DEBUG nova.compute.manager [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Refreshing instance network info cache due to event network-changed-768a74c4-85dd-4101-bd5a-11290b69e3a8. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.761945] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquiring lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.762075] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Acquired lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.762258] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Refreshing network info cache for port 768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.774907] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Releasing lock "refresh_cache-f3a6527c-777e-4a07-9482-598de15d4eb3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.776047] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 824.776598] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d77b4a2b-0a7b-4fa6-b9f4-c9f0a11f0cb4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.786349] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 824.786349] env[62000]: value = "task-882305" [ 824.786349] env[62000]: _type = "Task" [ 824.786349] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.806204] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882305, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.810343] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882304, 'name': Rename_Task, 'duration_secs': 0.233267} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 824.810736] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 824.810870] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dfa895a-eb8b-469a-a457-9a814cf447f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.820268] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 824.820268] env[62000]: value = "task-882306" [ 824.820268] env[62000]: _type = "Task" [ 824.820268] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.833302] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.847643] env[62000]: DEBUG nova.compute.manager [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 824.848016] env[62000]: DEBUG nova.compute.manager [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing instance network info cache due to event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 824.848197] env[62000]: DEBUG oslo_concurrency.lockutils [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.848358] env[62000]: DEBUG oslo_concurrency.lockutils [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.848561] env[62000]: DEBUG nova.network.neutron [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 824.919482] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 824.928120] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.929029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.929029] env[62000]: DEBUG nova.compute.manager [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 824.933164] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da417b0-ec98-4175-9fb8-63a1d9703ec1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.940073] env[62000]: DEBUG nova.compute.manager [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 824.940073] env[62000]: DEBUG nova.objects.instance [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'flavor' on Instance uuid e2e3270d-086b-4441-a3d6-49b05a60b51f {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 825.021420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.021420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.021420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.021420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.021420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.023692] env[62000]: INFO nova.compute.manager [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Terminating instance [ 825.024621] env[62000]: DEBUG nova.compute.manager [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.024946] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.025878] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ab5a352-5960-4bfd-8320-d751fb6ab5a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.037667] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.040623] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d97bf538-7a39-4170-9a57-d6959034bac2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.055557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.088710] env[62000]: INFO nova.compute.manager [-] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Took 1.05 seconds to deallocate network for instance. [ 825.181419] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.196150] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.196150] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.196150] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore2] ae1e8ebc-ddb8-4e95-847e-b9684e9161d4 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.196681] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-676e2225-2dd8-4a0f-82ad-0ebc4560e779 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.208936] env[62000]: DEBUG oslo_vmware.api [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 825.208936] env[62000]: value = "task-882308" [ 825.208936] env[62000]: _type = "Task" [ 825.208936] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.217703] env[62000]: DEBUG oslo_vmware.api [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882308, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.241581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.241581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.241899] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.242036] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 825.242249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.244533] env[62000]: INFO nova.compute.manager [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Terminating instance [ 825.247446] env[62000]: DEBUG nova.compute.manager [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 825.247691] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 825.248723] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b3b573-0ccc-4f61-ba57-ed1fa04bc20e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.261477] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.261714] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-861f8db4-72a4-4613-ae0e-c61705e46729 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.273633] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 825.273633] env[62000]: value = "task-882309" [ 825.273633] env[62000]: _type = "Task" [ 825.273633] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.287219] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882309, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.304129] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882305, 'name': PowerOffVM_Task, 'duration_secs': 0.258943} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.304421] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.305316] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:54:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='622b3a6c-bfbe-431a-b5c4-116aa70b0e48',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-862518541',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.305582] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.305753] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.305945] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.306171] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.306268] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.306478] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.306654] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.306815] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.306982] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.307184] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.312582] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4dc6a92-8875-458c-8e1b-fe3b197e6faa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.339779] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882306, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.341277] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 825.341277] env[62000]: value = "task-882310" [ 825.341277] env[62000]: _type = "Task" [ 825.341277] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.358144] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882310, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.408526] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55de4ab-782a-4f6b-a5cb-e969e0b4ce2b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.418435] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10da62b5-0455-46a5-8685-2ebbaa6cf3a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.459510] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 825.463043] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bd6ee9e-2937-40e6-8dea-969f5763b2b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.465608] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-13c8ee32-d047-4f70-a664-1fa538ffb235 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.474443] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578cbd96-a853-4b3f-a66f-d2200f1d9f01 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.480051] env[62000]: DEBUG oslo_vmware.api [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 825.480051] env[62000]: value = "task-882311" [ 825.480051] env[62000]: _type = "Task" [ 825.480051] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.497586] env[62000]: DEBUG nova.compute.provider_tree [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.505950] env[62000]: DEBUG oslo_vmware.api [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882311, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.539763] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Successfully created port: 7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 825.599459] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.717164] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.728066] env[62000]: DEBUG oslo_vmware.api [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882308, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.280392} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.728770] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.728770] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 825.728770] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 825.729032] env[62000]: INFO nova.compute.manager [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Took 0.70 seconds to destroy the instance on the hypervisor. [ 825.729654] env[62000]: DEBUG oslo.service.loopingcall [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.729654] env[62000]: DEBUG nova.compute.manager [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.729654] env[62000]: DEBUG nova.network.neutron [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 825.785143] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882309, 'name': PowerOffVM_Task, 'duration_secs': 0.263794} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.789829] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 825.789829] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 825.789829] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60160ce0-48b1-4c68-8f61-7efa6eeb94dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.844213] env[62000]: DEBUG oslo_vmware.api [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882306, 'name': PowerOnVM_Task, 'duration_secs': 0.601924} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.849158] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 825.849614] env[62000]: INFO nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Took 8.34 seconds to spawn the instance on the hypervisor. [ 825.849798] env[62000]: DEBUG nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 825.850631] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a856fde-5fd2-42e6-9adc-9007f440ffa0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.863556] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882310, 'name': ReconfigVM_Task, 'duration_secs': 0.255049} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.874137] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-066edb0f-eb3a-4a2f-a469-6f46610c6eb2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.904040] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:54:59Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='622b3a6c-bfbe-431a-b5c4-116aa70b0e48',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-862518541',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.904446] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.904606] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.904705] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.904836] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.905022] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.905260] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.905425] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.905614] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.905761] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.906188] env[62000]: DEBUG nova.virt.hardware [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.908555] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Updated VIF entry in instance network info cache for port 768a74c4-85dd-4101-bd5a-11290b69e3a8. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 825.912116] env[62000]: DEBUG nova.network.neutron [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Updating instance_info_cache with network_info: [{"id": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "address": "fa:16:3e:52:df:58", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap768a74c4-85", "ovs_interfaceid": "768a74c4-85dd-4101-bd5a-11290b69e3a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.912116] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9be6baf4-8bbd-47ba-abff-8b94d6aaab21 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.912913] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 825.913067] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 825.913307] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleting the datastore file [datastore1] 96107e36-bf3f-4ef0-8d8b-5c9601f4f514 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.913823] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c734995e-bc4d-4fb8-8d6c-32c22c6c4060 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.919655] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 825.919655] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520d7b3f-a8f2-67dc-dbf6-745a45142fe0" [ 825.919655] env[62000]: _type = "Task" [ 825.919655] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.926075] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 825.926075] env[62000]: value = "task-882313" [ 825.926075] env[62000]: _type = "Task" [ 825.926075] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.935021] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520d7b3f-a8f2-67dc-dbf6-745a45142fe0, 'name': SearchDatastore_Task, 'duration_secs': 0.009348} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.941091] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfiguring VM instance instance-00000032 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 825.944962] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c86e7309-8780-4c9f-9b8a-a64fe04adbd5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.959240] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882313, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.960192] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 825.967052] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 825.967052] env[62000]: value = "task-882314" [ 825.967052] env[62000]: _type = "Task" [ 825.967052] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.977900] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882314, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.989530] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 825.989793] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 825.989929] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 825.990134] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 825.990287] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 825.990438] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 825.990647] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 825.990809] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 825.991078] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 825.991258] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 825.991556] env[62000]: DEBUG nova.virt.hardware [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 825.992499] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df97be7f-9c86-436e-8d05-2f19956d329a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.002062] env[62000]: DEBUG nova.scheduler.client.report [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 826.009738] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eec0d994-f7a5-4cbe-be1d-ab3fd51e71a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.014196] env[62000]: DEBUG oslo_vmware.api [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882311, 'name': PowerOffVM_Task, 'duration_secs': 0.523235} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.017189] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 826.017395] env[62000]: DEBUG nova.compute.manager [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 826.020994] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9677044e-9195-4ccd-bea2-e1910413f419 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.391352] env[62000]: INFO nova.compute.manager [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Took 15.90 seconds to build instance. [ 826.414738] env[62000]: DEBUG oslo_concurrency.lockutils [req-35cb7208-c404-422e-a059-d6e6c9ba5b75 req-bf2eed37-b0f6-4f01-85f9-0d1faf4970b9 service nova] Releasing lock "refresh_cache-48d29490-c174-4702-a449-3bf10bc9207c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.437761] env[62000]: DEBUG oslo_vmware.api [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882313, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.231647} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.437761] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 826.437761] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 826.437761] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 826.437895] env[62000]: INFO nova.compute.manager [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Took 1.19 seconds to destroy the instance on the hypervisor. [ 826.438263] env[62000]: DEBUG oslo.service.loopingcall [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 826.438489] env[62000]: DEBUG nova.compute.manager [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 826.438590] env[62000]: DEBUG nova.network.neutron [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 826.477599] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882314, 'name': ReconfigVM_Task, 'duration_secs': 0.249667} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 826.477901] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfigured VM instance instance-00000032 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 826.478719] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-019a3a4c-6632-432e-89e0-492338fbf0e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.504649] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfiguring VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 826.505029] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-608acc5b-b59e-4e01-aee8-c5008c808291 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.521841] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.623s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.522477] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.525238] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.962s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.529679] env[62000]: INFO nova.compute.claims [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.537277] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 826.537277] env[62000]: value = "task-882315" [ 826.537277] env[62000]: _type = "Task" [ 826.537277] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 826.543572] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c79989a8-f264-4798-92a1-a400cc2ad739 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.615s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.550543] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882315, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 826.637044] env[62000]: DEBUG nova.network.neutron [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updated VIF entry in instance network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 826.637451] env[62000]: DEBUG nova.network.neutron [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}, {"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.894221] env[62000]: DEBUG oslo_concurrency.lockutils [None req-91653dd7-2d99-429e-90c4-749e33f7e2ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.420s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.032801] env[62000]: DEBUG nova.compute.utils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 827.035363] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 827.035561] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 827.050283] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882315, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.142317] env[62000]: DEBUG oslo_concurrency.lockutils [req-4bbe152f-8326-4300-af78-6a4e5fbd653f req-f1f9f989-f2a1-4a5c-9b27-d01a374dd5ab service nova] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.201587] env[62000]: DEBUG nova.policy [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfe15ae102584204b0816ce6e36afdfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cac5f0a5704d434082131155e107d190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.239678] env[62000]: DEBUG nova.objects.instance [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lazy-loading 'flavor' on Instance uuid ea8ae421-1d7f-4814-bc0f-90a3316ad028 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 827.364956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.365214] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.546025] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.563768] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882315, 'name': ReconfigVM_Task, 'duration_secs': 0.739517} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 827.564260] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Reconfigured VM instance instance-00000032 to attach disk [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3/f3a6527c-777e-4a07-9482-598de15d4eb3.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 827.565522] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e35c369-b2af-42c7-88a3-a26cbc3e1ddf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.595025] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e52bff-0927-4b31-82a4-8bd2f85e06b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.618555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7762833-73ae-486a-bfe6-da7f29ee0e81 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.646727] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0caa72f-9614-4cf4-94c7-828ae9380ed8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.655872] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 827.656357] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c5e39662-84e4-4880-857b-d6269958321b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.669668] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 827.669668] env[62000]: value = "task-882316" [ 827.669668] env[62000]: _type = "Task" [ 827.669668] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.685233] env[62000]: DEBUG nova.network.neutron [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.685538] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882316, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.709027] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.711020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.711020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.711020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 827.711020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.715631] env[62000]: INFO nova.compute.manager [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Terminating instance [ 827.719068] env[62000]: DEBUG nova.compute.manager [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.719564] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 827.722052] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3460f6c7-576b-46a2-9536-d2eecee46d10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.731918] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 827.733530] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-33db062d-dfb9-4879-8a71-24d0a682bd29 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.744883] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 827.744883] env[62000]: value = "task-882317" [ 827.744883] env[62000]: _type = "Task" [ 827.744883] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 827.745451] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.745757] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.762818] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882317, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 827.763976] env[62000]: DEBUG nova.network.neutron [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.846825] env[62000]: DEBUG nova.compute.manager [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Received event network-changed-849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 827.847100] env[62000]: DEBUG nova.compute.manager [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Refreshing instance network info cache due to event network-changed-849930d2-4200-4ec9-ac15-51c243a118bf. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 827.847758] env[62000]: DEBUG oslo_concurrency.lockutils [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] Acquiring lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 827.847758] env[62000]: DEBUG oslo_concurrency.lockutils [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] Acquired lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 827.847758] env[62000]: DEBUG nova.network.neutron [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Refreshing network info cache for port 849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 827.867897] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.019780] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d68d9e-59c3-42a2-8666-be5795378a4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.030846] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38845163-9b4d-4c93-a629-655dfd612560 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.075317] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e7796d6-501b-4f89-8c82-4f71545589cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.087482] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-754c9669-7d05-4ac1-911d-5a172321759a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.097426] env[62000]: DEBUG nova.compute.manager [req-93d70f8d-895a-4520-9bad-e22189cac445 req-71952766-107b-43b7-8b15-b33a312c42af service nova] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Received event network-vif-deleted-e3405b94-e374-4e74-8dc3-f1bb44464833 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 828.112503] env[62000]: DEBUG nova.compute.provider_tree [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 828.180642] env[62000]: DEBUG oslo_vmware.api [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882316, 'name': PowerOnVM_Task, 'duration_secs': 0.465401} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.180921] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 828.187007] env[62000]: INFO nova.compute.manager [-] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Took 1.75 seconds to deallocate network for instance. [ 828.262283] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882317, 'name': PowerOffVM_Task, 'duration_secs': 0.22505} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.262283] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 828.262283] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 828.262283] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5e21a17a-cbc5-4de1-bd43-324b984c427e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.268823] env[62000]: INFO nova.compute.manager [-] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Took 2.54 seconds to deallocate network for instance. [ 828.341658] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 828.342416] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 828.342716] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Deleting the datastore file [datastore1] 85f5f90a-9591-4393-9bef-2ad7e6a1a82f {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 828.343591] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a92893b8-b05e-4045-8e39-b6e9606fdc64 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.359220] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for the task: (returnval){ [ 828.359220] env[62000]: value = "task-882319" [ 828.359220] env[62000]: _type = "Task" [ 828.359220] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.371655] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.402765] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.578149] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.602113] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.602113] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.602113] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.602113] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.602113] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.602685] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.603051] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.603357] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.603690] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.603962] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.604317] env[62000]: DEBUG nova.virt.hardware [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.605467] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1226ede-42f9-496a-b22a-119e433165d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.615857] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Successfully created port: 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 828.618817] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47058d7e-971f-4c9f-8b9e-17670213327f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.624185] env[62000]: DEBUG nova.scheduler.client.report [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.690520] env[62000]: DEBUG nova.network.neutron [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updated VIF entry in instance network info cache for port 849930d2-4200-4ec9-ac15-51c243a118bf. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 828.690950] env[62000]: DEBUG nova.network.neutron [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.694422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.778244] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.868601] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "48d29490-c174-4702-a449-3bf10bc9207c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.868914] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.869149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "48d29490-c174-4702-a449-3bf10bc9207c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.869659] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.869659] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.875359] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882319, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.875868] env[62000]: INFO nova.compute.manager [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Terminating instance [ 828.877926] env[62000]: DEBUG nova.compute.manager [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 828.878155] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 828.878940] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1199354-4b14-46d0-96e0-ce674e712b48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.887734] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 828.888008] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e94df37c-01f3-4285-a2d4-d14b0139bbd2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.896359] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 828.896359] env[62000]: value = "task-882320" [ 828.896359] env[62000]: _type = "Task" [ 828.896359] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.908228] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882320, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.132171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.606s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 829.132709] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 829.137712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.082s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.139774] env[62000]: INFO nova.compute.claims [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 829.194047] env[62000]: INFO nova.compute.manager [None req-56496e7e-7eba-449b-8ea1-f80e01c817fc tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance to original state: 'active' [ 829.201590] env[62000]: DEBUG oslo_concurrency.lockutils [req-e8085290-4e10-4e7e-bca4-d1f13d944cb4 req-2851173e-1165-4fbc-a646-10f67bdb832e service nova] Releasing lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.380640] env[62000]: DEBUG oslo_vmware.api [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Task: {'id': task-882319, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.565838} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.380900] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 829.381107] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 829.381380] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 829.381570] env[62000]: INFO nova.compute.manager [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Took 1.66 seconds to destroy the instance on the hypervisor. [ 829.381820] env[62000]: DEBUG oslo.service.loopingcall [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.382102] env[62000]: DEBUG nova.compute.manager [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.382229] env[62000]: DEBUG nova.network.neutron [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.409282] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882320, 'name': PowerOffVM_Task, 'duration_secs': 0.313744} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.409561] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 829.409728] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 829.410103] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b0730e1-cec5-4e58-881b-bf94d735f4a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.509688] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 829.509688] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 829.509904] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore2] 48d29490-c174-4702-a449-3bf10bc9207c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 829.510643] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7bc1308a-f3a7-45c7-ba9e-c38c15e465cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.525962] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 829.525962] env[62000]: value = "task-882322" [ 829.525962] env[62000]: _type = "Task" [ 829.525962] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.542782] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.601263] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.601263] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.606748] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-656e8902-9da0-40c9-889a-8d1aab0f9266" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.606871] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-656e8902-9da0-40c9-889a-8d1aab0f9266" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.639761] env[62000]: DEBUG nova.compute.utils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 829.641551] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 829.641551] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 829.669583] env[62000]: DEBUG nova.network.neutron [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 829.754488] env[62000]: DEBUG nova.policy [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0d35b645d3fd45d89255a6ac663519d4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '393befb2bee84cac97b8e3e55e125837', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.820934] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Successfully updated port: 7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 830.005055] env[62000]: INFO nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Rebuilding instance [ 830.043256] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882322, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.068227] env[62000]: DEBUG nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 830.070517] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3a8371-f38f-4d16-975b-124ab9e263e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.104235] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 830.109582] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.111017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.111017] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd10ad11-0d79-4e50-98ed-c643dccbe5e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.136445] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a741549f-de39-4372-9577-de1fb2d75d2c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.167291] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 830.180038] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfiguring VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 830.183568] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cffac783-ad6e-4301-95ff-b366676042a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.206788] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 830.206788] env[62000]: value = "task-882323" [ 830.206788] env[62000]: _type = "Task" [ 830.206788] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.225534] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.323958] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.324591] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.324591] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.545772] env[62000]: DEBUG oslo_vmware.api [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882322, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.535027} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.548832] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 830.549535] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 830.549535] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 830.549535] env[62000]: INFO nova.compute.manager [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Took 1.67 seconds to destroy the instance on the hypervisor. [ 830.549730] env[62000]: DEBUG oslo.service.loopingcall [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 830.550130] env[62000]: DEBUG nova.compute.manager [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 830.550229] env[62000]: DEBUG nova.network.neutron [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 830.562036] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdb44f0c-8138-4437-9ead-6793553cfd67 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.571140] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a8eed8-e1f8-49d5-bb08-90f59ab04fcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.579826] env[62000]: DEBUG nova.network.neutron [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.606401] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 830.608054] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.608307] env[62000]: DEBUG nova.compute.manager [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Inject network info {{(pid=62000) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7219}} [ 830.608592] env[62000]: DEBUG nova.compute.manager [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] network_info to inject: |[{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _inject_network_info /opt/stack/nova/nova/compute/manager.py:7220}} [ 830.613376] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfiguring VM instance to set the machine id {{(pid=62000) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1802}} [ 830.613650] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-085c0637-25eb-4b47-917a-e01dae09f9b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.617968] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8411d69e-170a-412e-8cc2-78bd4a113443 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.621783] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3c56c00e-1856-457c-8f23-c425a153f172 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.644675] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1f4ca88-143f-404a-8fe4-2c0f0f3c91f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.656660] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Received event network-vif-deleted-7e374e87-d2b5-49ef-9516-991307309c6d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.656660] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Received event network-changed-3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 830.656660] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Refreshing instance network info cache due to event network-changed-3c405e0f-e997-4845-a292-10cad3943d50. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 830.656660] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquiring lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.656660] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquired lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.656660] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Refreshing network info cache for port 3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 830.657946] env[62000]: DEBUG oslo_vmware.api [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 830.657946] env[62000]: value = "task-882325" [ 830.657946] env[62000]: _type = "Task" [ 830.657946] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.658148] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 830.658148] env[62000]: value = "task-882324" [ 830.658148] env[62000]: _type = "Task" [ 830.658148] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.659136] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.682455] env[62000]: DEBUG nova.compute.provider_tree [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 830.695389] env[62000]: DEBUG oslo_vmware.api [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882325, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.695782] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 830.695984] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 830.696896] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bee454-ac47-4721-93dc-81a974a151a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.706450] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 830.706787] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-369b07ad-fc49-4a51-ba53-fdb7a368d9be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.718540] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.875889] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.879905] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Successfully created port: 860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 830.931324] env[62000]: DEBUG nova.network.neutron [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.959994] env[62000]: DEBUG nova.compute.manager [req-af7d7ec4-7512-4076-987b-1397284aa46b req-d7b670d5-02d4-4cdc-ac18-23cd8d9406b3 service nova] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Received event network-vif-deleted-d8f74268-0a99-4e0d-8042-52662b7e2c21 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 831.005767] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 831.006057] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 831.006260] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] e2e3270d-086b-4441-a3d6-49b05a60b51f {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 831.006529] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f5d6d43e-dd5d-406d-8703-d64fcabb4a65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.015875] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 831.015875] env[62000]: value = "task-882327" [ 831.015875] env[62000]: _type = "Task" [ 831.015875] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.025869] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.177209] env[62000]: DEBUG oslo_vmware.api [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882325, 'name': ReconfigVM_Task, 'duration_secs': 0.239905} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.177209] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a986257e-cc77-48ee-a604-9db4d7b7746a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Reconfigured VM instance to set the machine id {{(pid=62000) _set_machine_id /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1805}} [ 831.184458] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 831.188305] env[62000]: DEBUG nova.scheduler.client.report [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.224436] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.229484] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 831.229844] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 831.232805] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 831.233353] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 831.233353] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 831.233456] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 831.234104] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 831.234104] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 831.234189] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 831.234320] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 831.234527] env[62000]: DEBUG nova.virt.hardware [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 831.236200] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7ec288c-cfc1-4d75-9cf4-8aa35b2b2871 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.249066] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af00831f-4b95-41e1-b735-c7b35dfef15b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.437375] env[62000]: INFO nova.compute.manager [-] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Took 2.05 seconds to deallocate network for instance. [ 831.451837] env[62000]: DEBUG nova.network.neutron [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.526866] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882327, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.693959] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.557s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.694608] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 831.698340] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.099s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.698340] env[62000]: DEBUG nova.objects.instance [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lazy-loading 'resources' on Instance uuid 153b86d9-dc2e-463d-87fd-155ec23e2abb {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 831.704837] env[62000]: DEBUG nova.network.neutron [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.722652] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.735822] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Successfully updated port: 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.741130] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updated VIF entry in instance network info cache for port 3c405e0f-e997-4845-a292-10cad3943d50. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 831.741524] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating instance_info_cache with network_info: [{"id": "3c405e0f-e997-4845-a292-10cad3943d50", "address": "fa:16:3e:3f:0b:64", "network": {"id": "14ad3cf7-b997-4f08-8510-945d34e17eac", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-315454476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7d4b4854b30648d8ac861f08fbdb6da9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "97b68ed7-8461-4345-b064-96a1dde53a86", "external-id": "nsx-vlan-transportzone-140", "segmentation_id": 140, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3c405e0f-e9", "ovs_interfaceid": "3c405e0f-e997-4845-a292-10cad3943d50", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.942731] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.959014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.959014] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Instance network_info: |[{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 831.959014] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:d6:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e199318-132a-4975-92a1-61d761a06165', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 831.966770] env[62000]: DEBUG oslo.service.loopingcall [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 831.967037] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 831.967266] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c3398c5-a88d-46b2-ba27-93cd343acef0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.995178] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 831.995178] env[62000]: value = "task-882328" [ 831.995178] env[62000]: _type = "Task" [ 831.995178] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.004216] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882328, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.026259] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882327, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.573985} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.026531] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 832.026720] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 832.026895] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 832.202689] env[62000]: DEBUG nova.compute.utils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 832.209018] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 832.209451] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 832.211956] env[62000]: INFO nova.compute.manager [-] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Took 1.66 seconds to deallocate network for instance. [ 832.234514] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.240193] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.240331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.240801] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.248882] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Releasing lock "refresh_cache-2f35dab3-6c32-4a35-91ca-1d6859e27f2c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 832.248882] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Received event network-vif-plugged-7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.248882] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.248882] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.248882] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.249168] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] No waiting events found dispatching network-vif-plugged-7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 832.249398] env[62000]: WARNING nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Received unexpected event network-vif-plugged-7e199318-132a-4975-92a1-61d761a06165 for instance with vm_state building and task_state spawning. [ 832.250024] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 832.250024] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing instance network info cache due to event network-changed-c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 832.250024] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquiring lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.250277] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquired lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.250956] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Refreshing network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 832.298365] env[62000]: DEBUG nova.policy [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98f2c4bd042c488c91cd858883a2cea2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ed0fd1dd9fd74dbebdf9e919ae5da0d0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 832.508231] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882328, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.564881] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1aad24-a822-4b7a-ac31-26866732f851 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.574570] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de6a2946-48f4-4258-bcae-5686336a7f3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.611934] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.613248] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.613248] env[62000]: DEBUG nova.compute.manager [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 832.613959] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0db01e4f-41e4-42c8-8219-cf453242d64f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.616933] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10871db9-0e5f-4dcf-8ee6-55485ecfb896 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.626962] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b35c4f95-0a57-414c-b338-9ccb9c662108 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.631637] env[62000]: DEBUG nova.compute.manager [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 832.631637] env[62000]: DEBUG nova.objects.instance [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 832.644543] env[62000]: DEBUG nova.compute.provider_tree [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.710468] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 832.729925] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.732048] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.764394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "f3a6527c-777e-4a07-9482-598de15d4eb3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.765057] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.765399] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.765653] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.765888] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.769331] env[62000]: INFO nova.compute.manager [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Terminating instance [ 832.771394] env[62000]: DEBUG nova.compute.manager [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.771681] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 832.776392] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c1eb80-624f-449d-aeed-12704aa64640 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.787743] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 832.789283] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c4fdc497-0306-4b05-bc42-8688ed89b6dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.794706] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.794998] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.802863] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 832.802863] env[62000]: value = "task-882329" [ 832.802863] env[62000]: _type = "Task" [ 832.802863] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.813604] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.823098] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 832.834501] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Successfully created port: 7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.847844] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.848224] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.848456] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.848643] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.848815] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.851448] env[62000]: INFO nova.compute.manager [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Terminating instance [ 832.853581] env[62000]: DEBUG nova.compute.manager [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 832.853846] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 832.857653] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0dc0a466-561e-4d7f-b2b3-7b70b5566b1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.866283] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 832.866613] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3bc8523c-051e-4600-b572-7be53c3d0a7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.876846] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 832.876846] env[62000]: value = "task-882330" [ 832.876846] env[62000]: _type = "Task" [ 832.876846] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.887110] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.980208] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Successfully updated port: 860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 833.005665] env[62000]: DEBUG nova.compute.manager [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.005963] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.006220] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.006945] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.006945] env[62000]: DEBUG nova.compute.manager [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] No waiting events found dispatching network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.006945] env[62000]: WARNING nova.compute.manager [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received unexpected event network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 for instance with vm_state building and task_state spawning. [ 833.006945] env[62000]: DEBUG nova.compute.manager [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.007149] env[62000]: DEBUG nova.compute.manager [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing instance network info cache due to event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 833.007219] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.014401] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882328, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.062833] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.063154] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.063349] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.063575] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.063778] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.063986] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.064401] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.064656] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.064883] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.065111] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.065336] env[62000]: DEBUG nova.virt.hardware [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.066572] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93b1523-5f14-4f32-9ed0-c5a4b90bb89c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.077244] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5de0def2-f4b0-4ffd-bf26-aaebbb40b28d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.099148] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:c1:8c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0cb92206-16ee-49e1-8e70-fa3f39d678b4', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.109250] env[62000]: DEBUG oslo.service.loopingcall [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.109250] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.109250] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bbef4f20-3a69-44c4-98fb-14552a8af2af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.132215] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.132215] env[62000]: value = "task-882331" [ 833.132215] env[62000]: _type = "Task" [ 833.132215] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.141031] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 833.141390] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a9ece032-4d75-456d-8f7c-a27390cab393 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.146134] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882331, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.147841] env[62000]: DEBUG nova.scheduler.client.report [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.155986] env[62000]: DEBUG oslo_vmware.api [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 833.155986] env[62000]: value = "task-882332" [ 833.155986] env[62000]: _type = "Task" [ 833.155986] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.169717] env[62000]: DEBUG oslo_vmware.api [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882332, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.213532] env[62000]: DEBUG nova.network.neutron [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.240213] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.249864] env[62000]: DEBUG nova.compute.manager [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Received event network-vif-deleted-768a74c4-85dd-4101-bd5a-11290b69e3a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.250070] env[62000]: DEBUG nova.compute.manager [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Received event network-vif-plugged-860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.250267] env[62000]: DEBUG oslo_concurrency.lockutils [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] Acquiring lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.250475] env[62000]: DEBUG oslo_concurrency.lockutils [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.250644] env[62000]: DEBUG oslo_concurrency.lockutils [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.250900] env[62000]: DEBUG nova.compute.manager [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] No waiting events found dispatching network-vif-plugged-860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 833.250989] env[62000]: WARNING nova.compute.manager [req-859dea4a-a381-4aff-92ae-f45bfcf8e264 req-526cf35d-8201-43be-893f-80f4d8022ee1 service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Received unexpected event network-vif-plugged-860a5cc9-5273-47c1-a9e7-b173fb60c2a1 for instance with vm_state building and task_state spawning. [ 833.297461] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 833.314076] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882329, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.390928] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882330, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.430024] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updated VIF entry in instance network info cache for port c4e0acd0-a5a2-4120-86ff-826f7dbe160e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 833.430024] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [{"id": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "address": "fa:16:3e:e0:5d:f5", "network": {"id": "16bd471a-8387-404f-9986-63ceba14dbe7", "bridge": "br-int", "label": "tempest-AttachInterfacesUnderV243Test-1132644585-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.175", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0b5041eab4c4480c9221aefa7838ab27", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "674802e7-b847-4bef-a7a8-f90ac7a3a0a7", "external-id": "nsx-vlan-transportzone-953", "segmentation_id": 953, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc4e0acd0-a5", "ovs_interfaceid": "c4e0acd0-a5a2-4120-86ff-826f7dbe160e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.484768] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.484987] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquired lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.485147] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 833.511431] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882328, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.643686] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882331, 'name': CreateVM_Task, 'duration_secs': 0.361533} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.643860] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 833.644738] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.644738] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.645080] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 833.645344] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d2848a00-7d66-4a47-beab-43158ecc7133 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.650694] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 833.650694] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c59d89-63f7-e683-f9bd-f20cd8cb0b23" [ 833.650694] env[62000]: _type = "Task" [ 833.650694] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.655139] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.957s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.659969] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.943s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.661563] env[62000]: INFO nova.compute.claims [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.664722] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c59d89-63f7-e683-f9bd-f20cd8cb0b23, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.675572] env[62000]: DEBUG oslo_vmware.api [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882332, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.688328] env[62000]: INFO nova.scheduler.client.report [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Deleted allocations for instance 153b86d9-dc2e-463d-87fd-155ec23e2abb [ 833.719581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.720398] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance network_info: |[{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.723902] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.724122] env[62000]: DEBUG nova.network.neutron [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 833.729027] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:7d:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.733430] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating folder: Project (cac5f0a5704d434082131155e107d190). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.736304] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 833.737183] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6d579fa-c459-4269-ae91-480b94f70b84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.747295] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.751655] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created folder: Project (cac5f0a5704d434082131155e107d190) in parent group-v201431. [ 833.751859] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating folder: Instances. Parent ref: group-v201563. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 833.752160] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-79dbbd34-169e-4dd0-b30f-87b41d9027e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.763894] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 833.764167] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 833.764334] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 833.764658] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 833.764658] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 833.764811] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 833.765071] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 833.765198] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 833.765523] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 833.765523] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 833.765728] env[62000]: DEBUG nova.virt.hardware [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 833.766851] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99fdc9d7-0197-425c-b9d0-627ecad4cdf4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.770716] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created folder: Instances in parent group-v201563. [ 833.771363] env[62000]: DEBUG oslo.service.loopingcall [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.771928] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 833.772190] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53d26887-3dc5-4386-bd98-1c92787dbc12 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.789782] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f507e1a-ae9b-45db-99a9-6572e3e36948 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.794449] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.794449] env[62000]: value = "task-882335" [ 833.794449] env[62000]: _type = "Task" [ 833.794449] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.814904] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882335, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.820439] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882329, 'name': PowerOffVM_Task, 'duration_secs': 0.883587} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.821053] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 833.821253] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 833.821507] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1de847b1-1c75-4bcf-8b2d-3f6d1e8cd16d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.825482] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.888400] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882330, 'name': PowerOffVM_Task, 'duration_secs': 0.841214} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.888742] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 833.888945] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 833.889213] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-747edd91-d1f5-4c65-af82-a56ee71ca393 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.931613] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Releasing lock "refresh_cache-ea8ae421-1d7f-4814-bc0f-90a3316ad028" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.931903] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Received event network-changed-7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 833.932166] env[62000]: DEBUG nova.compute.manager [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Refreshing instance network info cache due to event network-changed-7e199318-132a-4975-92a1-61d761a06165. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 833.932391] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquiring lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.932537] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Acquired lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.932700] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Refreshing network info cache for port 7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 834.010746] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882328, 'name': CreateVM_Task, 'duration_secs': 1.525852} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.010876] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.011566] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.018958] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.161849] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c59d89-63f7-e683-f9bd-f20cd8cb0b23, 'name': SearchDatastore_Task, 'duration_secs': 0.011011} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.162226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.162474] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.162714] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.162865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.163095] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.163386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.163690] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.163987] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a91c9865-aa6f-4d3a-adb7-527b4e60ade0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.165697] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fddeb7f9-b4ec-4480-bda1-3c5f63d3b7e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.180950] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 834.180950] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f9bd73-8440-4191-0bc7-c74e2bd9d3ef" [ 834.180950] env[62000]: _type = "Task" [ 834.180950] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.191101] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.191345] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 834.192189] env[62000]: DEBUG oslo_vmware.api [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882332, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.192480] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7760e8dc-3e0c-4bdc-b4f2-a5e4730c68ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.201685] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f9bd73-8440-4191-0bc7-c74e2bd9d3ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.202233] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cbbd071a-2ade-4314-b2bf-26b62b634921 tempest-ServerShowV254Test-226654429 tempest-ServerShowV254Test-226654429-project-member] Lock "153b86d9-dc2e-463d-87fd-155ec23e2abb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.411s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.204920] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 834.204920] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ba70f9-f552-f617-1a34-7062e1563092" [ 834.204920] env[62000]: _type = "Task" [ 834.204920] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.214315] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ba70f9-f552-f617-1a34-7062e1563092, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.229361] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.230299] env[62000]: DEBUG nova.network.neutron [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Updating instance_info_cache with network_info: [{"id": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "address": "fa:16:3e:30:55:59", "network": {"id": "3f5a6ad5-6cfd-46e6-ae3f-4b3c4c874a73", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1004085576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "393befb2bee84cac97b8e3e55e125837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23fc30ea-1f06-424d-86e1-27ae5435b1a9", "external-id": "nsx-vlan-transportzone-189", "segmentation_id": 189, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860a5cc9-52", "ovs_interfaceid": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.298158] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.298158] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.298158] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Deleting the datastore file [datastore1] f3a6527c-777e-4a07-9482-598de15d4eb3 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.299893] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1599826b-c640-4b06-b526-a1732cf181d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.309123] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882335, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.311116] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for the task: (returnval){ [ 834.311116] env[62000]: value = "task-882338" [ 834.311116] env[62000]: _type = "Task" [ 834.311116] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.320226] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.324412] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 834.324621] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 834.327047] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Deleting the datastore file [datastore1] ea8ae421-1d7f-4814-bc0f-90a3316ad028 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 834.327047] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f6ffc5e5-c27d-47dc-bf6c-088ab1636efd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.332488] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for the task: (returnval){ [ 834.332488] env[62000]: value = "task-882339" [ 834.332488] env[62000]: _type = "Task" [ 834.332488] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.342668] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882339, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.681212] env[62000]: DEBUG oslo_vmware.api [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882332, 'name': PowerOffVM_Task, 'duration_secs': 1.225106} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.681461] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 834.681650] env[62000]: DEBUG nova.compute.manager [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.682600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90e12eda-e8fc-4879-81f0-f208d3c6ffaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.689039] env[62000]: DEBUG nova.network.neutron [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updated VIF entry in instance network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 834.689395] env[62000]: DEBUG nova.network.neutron [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.722553] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f9bd73-8440-4191-0bc7-c74e2bd9d3ef, 'name': SearchDatastore_Task, 'duration_secs': 0.025383} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.731355] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.731355] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.731524] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.739655] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Releasing lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.739655] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Instance network_info: |[{"id": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "address": "fa:16:3e:30:55:59", "network": {"id": "3f5a6ad5-6cfd-46e6-ae3f-4b3c4c874a73", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1004085576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "393befb2bee84cac97b8e3e55e125837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23fc30ea-1f06-424d-86e1-27ae5435b1a9", "external-id": "nsx-vlan-transportzone-189", "segmentation_id": 189, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860a5cc9-52", "ovs_interfaceid": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 834.742209] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:30:55:59', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '23fc30ea-1f06-424d-86e1-27ae5435b1a9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '860a5cc9-5273-47c1-a9e7-b173fb60c2a1', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 834.749405] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Creating folder: Project (393befb2bee84cac97b8e3e55e125837). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.749720] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ba70f9-f552-f617-1a34-7062e1563092, 'name': SearchDatastore_Task, 'duration_secs': 0.016464} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.750663] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-626db369-b2a0-4c32-a197-d42cb1951d17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.758509] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55ea2c25-ff89-4288-b6d1-136eaee9fab0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.760740] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.765150] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 834.765150] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52156a83-cdea-3e02-0586-cdd66844d837" [ 834.765150] env[62000]: _type = "Task" [ 834.765150] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.773444] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52156a83-cdea-3e02-0586-cdd66844d837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.774792] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Created folder: Project (393befb2bee84cac97b8e3e55e125837) in parent group-v201431. [ 834.774973] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Creating folder: Instances. Parent ref: group-v201566. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 834.775231] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dff3b736-ec72-4d4f-acb3-b28da76204c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.789779] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Created folder: Instances in parent group-v201566. [ 834.790091] env[62000]: DEBUG oslo.service.loopingcall [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.790295] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 834.792761] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5f2a6c6f-eaad-41ec-a9f3-ba3f84ddcc31 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.821436] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 834.821436] env[62000]: value = "task-882342" [ 834.821436] env[62000]: _type = "Task" [ 834.821436] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.828860] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882335, 'name': CreateVM_Task, 'duration_secs': 0.706842} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.829658] env[62000]: DEBUG oslo_vmware.api [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Task: {'id': task-882338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377928} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.835881] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 834.836192] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.836388] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.836568] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.836747] env[62000]: INFO nova.compute.manager [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Took 2.07 seconds to destroy the instance on the hypervisor. [ 834.836989] env[62000]: DEBUG oslo.service.loopingcall [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.838425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.838425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.838759] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.838759] env[62000]: DEBUG nova.compute.manager [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.838759] env[62000]: DEBUG nova.network.neutron [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.843382] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0711445-bd30-4c7c-89b1-dfe6345eb1bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.848874] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882342, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.857252] env[62000]: DEBUG oslo_vmware.api [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Task: {'id': task-882339, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.344521} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.857252] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 834.857252] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5258791c-0bdc-14e3-51e1-2e79e1e2565f" [ 834.857252] env[62000]: _type = "Task" [ 834.857252] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.857252] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 834.857252] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 834.857252] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 834.857252] env[62000]: INFO nova.compute.manager [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Took 2.00 seconds to destroy the instance on the hypervisor. [ 834.857252] env[62000]: DEBUG oslo.service.loopingcall [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.857252] env[62000]: DEBUG nova.compute.manager [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.857252] env[62000]: DEBUG nova.network.neutron [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 834.867664] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5258791c-0bdc-14e3-51e1-2e79e1e2565f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.086822] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3444cddb-8371-4860-9683-8dd761a04b66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.098540] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122f0a47-9ca3-4e24-8aa7-d2096ad801d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.132635] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff67e983-48f8-40a2-a670-e6a251c56cb8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.141037] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cff5969b-31fa-4b10-bc51-6995c075ede1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.155574] env[62000]: DEBUG nova.compute.provider_tree [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 835.207730] env[62000]: DEBUG oslo_concurrency.lockutils [req-06b79c27-a7bc-4fa6-856c-4bedfcf3b619 req-75281d99-f1c0-4ea0-b4c7-876f5c0cc962 service nova] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.236515] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3484f87e-e9ce-49cf-8b8f-d1631a0b2276 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 2.624s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.246873] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.276600] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52156a83-cdea-3e02-0586-cdd66844d837, 'name': SearchDatastore_Task, 'duration_secs': 0.021698} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.276906] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.277388] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 835.277542] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.277767] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 835.278032] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-131f2a29-6dc1-49b9-bc8d-43df6bc86cfc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.281050] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8386b6ec-3286-4079-8342-71ada583dbcb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.289613] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 835.289613] env[62000]: value = "task-882343" [ 835.289613] env[62000]: _type = "Task" [ 835.289613] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.290974] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 835.291139] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 835.296270] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2553b535-8275-42aa-ae45-1e562195477e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.303417] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882343, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.304588] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 835.304588] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52624f46-c9ed-c3d4-50fc-4b6ea2e2e42d" [ 835.304588] env[62000]: _type = "Task" [ 835.304588] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.312848] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52624f46-c9ed-c3d4-50fc-4b6ea2e2e42d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.340140] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882342, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.370391] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5258791c-0bdc-14e3-51e1-2e79e1e2565f, 'name': SearchDatastore_Task, 'duration_secs': 0.016085} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.370775] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.371304] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 835.371594] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.413060] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updated VIF entry in instance network info cache for port 7e199318-132a-4975-92a1-61d761a06165. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 835.413611] env[62000]: DEBUG nova.network.neutron [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.519299] env[62000]: DEBUG nova.compute.manager [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Received event network-changed-860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.519425] env[62000]: DEBUG nova.compute.manager [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Refreshing instance network info cache due to event network-changed-860a5cc9-5273-47c1-a9e7-b173fb60c2a1. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 835.519673] env[62000]: DEBUG oslo_concurrency.lockutils [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] Acquiring lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.519940] env[62000]: DEBUG oslo_concurrency.lockutils [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] Acquired lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.520336] env[62000]: DEBUG nova.network.neutron [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Refreshing network info cache for port 860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 835.632644] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Successfully updated port: 7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 835.660664] env[62000]: DEBUG nova.scheduler.client.report [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.748363] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.757459] env[62000]: DEBUG nova.compute.manager [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Received event network-vif-plugged-7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 835.757687] env[62000]: DEBUG oslo_concurrency.lockutils [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] Acquiring lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.758028] env[62000]: DEBUG oslo_concurrency.lockutils [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.758263] env[62000]: DEBUG oslo_concurrency.lockutils [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.758548] env[62000]: DEBUG nova.compute.manager [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] No waiting events found dispatching network-vif-plugged-7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.758638] env[62000]: WARNING nova.compute.manager [req-11439eed-8952-484f-ba88-016f948d43aa req-96095586-51b4-49ba-9a21-a1496382bbbe service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Received unexpected event network-vif-plugged-7801caca-efe4-49ff-b583-7e91be8cf888 for instance with vm_state building and task_state spawning. [ 835.801491] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882343, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.816756] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52624f46-c9ed-c3d4-50fc-4b6ea2e2e42d, 'name': SearchDatastore_Task, 'duration_secs': 0.012758} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.817671] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c5e0006-39e2-4bc1-8199-a8313496117a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.825503] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 835.825503] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fba358-efbc-04e1-5754-3af6e6d202ae" [ 835.825503] env[62000]: _type = "Task" [ 835.825503] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.839010] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fba358-efbc-04e1-5754-3af6e6d202ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.842610] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882342, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.918647] env[62000]: DEBUG oslo_concurrency.lockutils [req-f02a7932-34b3-495b-a814-c2238cce1077 req-e6bd294c-dd14-4419-a8b0-1ba7d2f4e2af service nova] Releasing lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.051227] env[62000]: DEBUG nova.network.neutron [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.135238] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.135431] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquired lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.135594] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 836.148902] env[62000]: DEBUG nova.network.neutron [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.168031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.508s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.168656] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 836.171614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.770s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.174308] env[62000]: INFO nova.compute.claims [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.244770] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.303400] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882343, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.655807} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.303676] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 836.303894] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 836.304481] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-45cd370e-87ea-4b29-9ed1-f05f5cbfc1aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.311673] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 836.311673] env[62000]: value = "task-882344" [ 836.311673] env[62000]: _type = "Task" [ 836.311673] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.319870] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882344, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.335436] env[62000]: DEBUG nova.network.neutron [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Updated VIF entry in instance network info cache for port 860a5cc9-5273-47c1-a9e7-b173fb60c2a1. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 836.335436] env[62000]: DEBUG nova.network.neutron [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Updating instance_info_cache with network_info: [{"id": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "address": "fa:16:3e:30:55:59", "network": {"id": "3f5a6ad5-6cfd-46e6-ae3f-4b3c4c874a73", "bridge": "br-int", "label": "tempest-ImagesNegativeTestJSON-1004085576-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "393befb2bee84cac97b8e3e55e125837", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "23fc30ea-1f06-424d-86e1-27ae5435b1a9", "external-id": "nsx-vlan-transportzone-189", "segmentation_id": 189, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap860a5cc9-52", "ovs_interfaceid": "860a5cc9-5273-47c1-a9e7-b173fb60c2a1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.338995] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fba358-efbc-04e1-5754-3af6e6d202ae, 'name': SearchDatastore_Task, 'duration_secs': 0.050327} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.339814] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.340096] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.340371] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.340563] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.340772] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-845bfd02-7640-4ee9-b24c-e5cf6b10d774 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.345645] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-028f2b85-7492-4b7c-b9dc-8ed689a8562d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.347228] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882342, 'name': CreateVM_Task, 'duration_secs': 1.364675} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.348197] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 836.349083] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.349254] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.349569] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.350107] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-94d4f615-1531-4edd-aae3-674b0a2d3e74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.353381] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 836.353381] env[62000]: value = "task-882345" [ 836.353381] env[62000]: _type = "Task" [ 836.353381] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.358595] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 836.358595] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5292f1f5-28e5-8b95-a484-4ae0bfcc4af9" [ 836.358595] env[62000]: _type = "Task" [ 836.358595] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.358827] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.358995] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.360090] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1973a2aa-5dfd-4ca0-9c7d-c795be7773b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.369006] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882345, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.372332] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 836.372332] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f39ac3-d993-7277-1144-cd9056880fb7" [ 836.372332] env[62000]: _type = "Task" [ 836.372332] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.375731] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5292f1f5-28e5-8b95-a484-4ae0bfcc4af9, 'name': SearchDatastore_Task, 'duration_secs': 0.009327} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.378667] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.378904] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 836.379133] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.384650] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f39ac3-d993-7277-1144-cd9056880fb7, 'name': SearchDatastore_Task, 'duration_secs': 0.008936} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.385383] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca64e048-f880-48ff-a51d-e14d91c1ef47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.390747] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 836.390747] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c29b2a-bde9-4c0e-f71d-d0c143d755ee" [ 836.390747] env[62000]: _type = "Task" [ 836.390747] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.397951] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c29b2a-bde9-4c0e-f71d-d0c143d755ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.557109] env[62000]: INFO nova.compute.manager [-] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Took 1.70 seconds to deallocate network for instance. [ 836.653269] env[62000]: INFO nova.compute.manager [-] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Took 1.81 seconds to deallocate network for instance. [ 836.680637] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.683922] env[62000]: DEBUG nova.compute.utils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.687961] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.688147] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 836.747083] env[62000]: DEBUG oslo_vmware.api [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882323, 'name': ReconfigVM_Task, 'duration_secs': 6.367577} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.747367] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.747910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Reconfigured VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 836.771325] env[62000]: DEBUG nova.policy [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d0b65722357449ab64736e5e1eb713f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ce3270692934d2c9c1330a45c0e059e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.823812] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882344, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071823} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.824369] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.825343] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c8320d-e822-4537-a066-fae38f8525cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.842994] env[62000]: DEBUG oslo_concurrency.lockutils [req-474f3196-56a6-4e16-bc7d-f94ecf04d06a req-8cf1af40-643a-43e9-b7a4-bde0b2aa35ae service nova] Releasing lock "refresh_cache-e9f2cd53-2c73-4c1f-858a-98a291da5547" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.852400] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.852770] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-258a40c9-2cd7-49f3-a394-7199cef8f32f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.882939] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882345, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.885142] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 836.885142] env[62000]: value = "task-882346" [ 836.885142] env[62000]: _type = "Task" [ 836.885142] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.898332] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882346, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.905103] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c29b2a-bde9-4c0e-f71d-d0c143d755ee, 'name': SearchDatastore_Task, 'duration_secs': 0.007942} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.905473] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.905748] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 836.906053] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.906253] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 836.906485] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ad57f895-a4e0-4762-9481-6fb9efd29a38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.908768] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4dca5c54-cee0-438e-9d12-a446737f0531 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.918289] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 836.918289] env[62000]: value = "task-882347" [ 836.918289] env[62000]: _type = "Task" [ 836.918289] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.923737] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 836.923960] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 836.928169] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71b38ea9-56a0-434a-8adf-ef2a49153ca1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.933582] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882347, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.938783] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 836.938783] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52020216-8c31-8a8e-5758-69276329d294" [ 836.938783] env[62000]: _type = "Task" [ 836.938783] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.947744] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52020216-8c31-8a8e-5758-69276329d294, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.961154] env[62000]: DEBUG nova.network.neutron [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updating instance_info_cache with network_info: [{"id": "7801caca-efe4-49ff-b583-7e91be8cf888", "address": "fa:16:3e:c3:e3:98", "network": {"id": "51b6593d-5326-4a3d-9dfd-befc01ae8135", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-2013827846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed0fd1dd9fd74dbebdf9e919ae5da0d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7801caca-ef", "ovs_interfaceid": "7801caca-efe4-49ff-b583-7e91be8cf888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.967834] env[62000]: DEBUG nova.compute.manager [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Stashing vm_state: stopped {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 837.787080] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Successfully created port: 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 837.792287] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.793216] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.795898] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 837.798808] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Releasing lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.799434] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Instance network_info: |[{"id": "7801caca-efe4-49ff-b583-7e91be8cf888", "address": "fa:16:3e:c3:e3:98", "network": {"id": "51b6593d-5326-4a3d-9dfd-befc01ae8135", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-2013827846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed0fd1dd9fd74dbebdf9e919ae5da0d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7801caca-ef", "ovs_interfaceid": "7801caca-efe4-49ff-b583-7e91be8cf888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.807017] env[62000]: DEBUG nova.compute.manager [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Received event network-vif-deleted-c4e0acd0-a5a2-4120-86ff-826f7dbe160e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.807017] env[62000]: DEBUG nova.compute.manager [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-deleted-656e8902-9da0-40c9-889a-8d1aab0f9266 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.807017] env[62000]: INFO nova.compute.manager [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Neutron deleted interface 656e8902-9da0-40c9-889a-8d1aab0f9266; detaching it from the instance and deleting it from the info cache [ 837.807017] env[62000]: DEBUG nova.network.neutron [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "515167d1-0a88-4083-b014-249c351061a8", "address": "fa:16:3e:16:52:5d", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap515167d1-0a", "ovs_interfaceid": "515167d1-0a88-4083-b014-249c351061a8", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.817286] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c3:e3:98', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '39a4aca0-934b-4a91-8779-6a4360c3f967', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7801caca-efe4-49ff-b583-7e91be8cf888', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.824697] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Creating folder: Project (ed0fd1dd9fd74dbebdf9e919ae5da0d0). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.827479] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bd001185-c8a8-4980-95c8-7edcbe07c1a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.831295] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.845694] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882345, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744539} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.845769] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882346, 'name': ReconfigVM_Task, 'duration_secs': 0.783621} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.847160] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52020216-8c31-8a8e-5758-69276329d294, 'name': SearchDatastore_Task, 'duration_secs': 0.059908} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.847389] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882347, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.67193} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.851571] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.851814] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.855018] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Reconfigured VM instance instance-0000003f to attach disk [datastore2] e2e3270d-086b-4441-a3d6-49b05a60b51f/e2e3270d-086b-4441-a3d6-49b05a60b51f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.855018] env[62000]: DEBUG nova.compute.manager [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Received event network-changed-7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 837.855018] env[62000]: DEBUG nova.compute.manager [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Refreshing instance network info cache due to event network-changed-7801caca-efe4-49ff-b583-7e91be8cf888. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 837.855018] env[62000]: DEBUG oslo_concurrency.lockutils [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] Acquiring lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.855018] env[62000]: DEBUG oslo_concurrency.lockutils [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] Acquired lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.855018] env[62000]: DEBUG nova.network.neutron [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Refreshing network info cache for port 7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 837.856989] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 837.857203] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 837.858813] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Created folder: Project (ed0fd1dd9fd74dbebdf9e919ae5da0d0) in parent group-v201431. [ 837.858986] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Creating folder: Instances. Parent ref: group-v201569. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 837.859431] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-033aee0e-5982-4221-9f11-c62d12214d04 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.861388] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9e32609e-e9b0-489c-8288-1c95383c0c92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.863449] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99d6fd01-be06-43a4-aa37-2a9dcee5eae8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.865453] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb53d54b-14e7-4459-a58e-7f8fa2c9a92a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.867470] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a2c8aafe-8c96-4f07-8015-f703088aa450 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.879509] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 837.879509] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f34e52-fa06-e747-9946-e3d0def2ece0" [ 837.879509] env[62000]: _type = "Task" [ 837.879509] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.880091] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 837.880091] env[62000]: value = "task-882350" [ 837.880091] env[62000]: _type = "Task" [ 837.880091] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.880382] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 837.880382] env[62000]: value = "task-882351" [ 837.880382] env[62000]: _type = "Task" [ 837.880382] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.881675] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 837.881675] env[62000]: value = "task-882349" [ 837.881675] env[62000]: _type = "Task" [ 837.881675] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.897746] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Created folder: Instances in parent group-v201569. [ 837.898641] env[62000]: DEBUG oslo.service.loopingcall [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.899412] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 837.900068] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ae14a64-1f39-4d85-95d2-ca2f26270fea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.929098] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882349, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.929497] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882351, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.929696] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f34e52-fa06-e747-9946-e3d0def2ece0, 'name': SearchDatastore_Task, 'duration_secs': 0.025599} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.929918] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882350, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.930746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.931176] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] e9f2cd53-2c73-4c1f-858a-98a291da5547/e9f2cd53-2c73-4c1f-858a-98a291da5547.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 837.931428] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c7fc7efc-37cc-492f-90f4-50e233b07f33 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.936883] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.936883] env[62000]: value = "task-882353" [ 837.936883] env[62000]: _type = "Task" [ 837.936883] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.943817] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 837.943817] env[62000]: value = "task-882354" [ 837.943817] env[62000]: _type = "Task" [ 837.943817] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.953148] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882353, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.957888] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882354, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.216645] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-665a6bf4-b099-4bd9-bf06-ce39540a2eb7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.225618] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160dd8b0-5711-4d20-91d5-538c187830ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.262487] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1b0a47-e954-473e-a0af-0f05a5e03079 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.280588] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50fff1ce-48f0-4ade-9db5-2deb7e18cf9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.299879] env[62000]: DEBUG nova.compute.provider_tree [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.332797] env[62000]: DEBUG oslo_concurrency.lockutils [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.333094] env[62000]: DEBUG oslo_concurrency.lockutils [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] Acquired lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.334146] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eacfea3-ea52-453d-aa5c-a58497e8899f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.361321] env[62000]: DEBUG oslo_concurrency.lockutils [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] Releasing lock "71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.361652] env[62000]: WARNING nova.compute.manager [req-05ef2585-5513-4c93-b25b-a47d20b4702a req-8eb7efa1-03cd-47e4-ae6f-c69fc763b5fe service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Detach interface failed, port_id=656e8902-9da0-40c9-889a-8d1aab0f9266, reason: No device with interface-id 656e8902-9da0-40c9-889a-8d1aab0f9266 exists on VM: nova.exception.NotFound: No device with interface-id 656e8902-9da0-40c9-889a-8d1aab0f9266 exists on VM [ 838.399029] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882351, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151998} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.405591] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.406505] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882349, 'name': Rename_Task, 'duration_secs': 0.250557} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.406732] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882350, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.137496} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.410786] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4754437-d9f3-4c43-881e-a726dd550e5f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.413025] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 838.413462] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 838.414096] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-49b795c0-e7f9-43b2-9f62-27a380df40ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.416444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246bcd14-764b-4e0c-a3c9-145a959aafd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.443298] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.454770] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-57e442a4-6f1b-42d6-9528-175a5d96c285 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.471454] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 838.471454] env[62000]: value = "task-882355" [ 838.471454] env[62000]: _type = "Task" [ 838.471454] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.480585] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 838.488525] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c6aef5d-eb76-4cea-81d2-37a03625bc50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.517431] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 838.517431] env[62000]: value = "task-882356" [ 838.517431] env[62000]: _type = "Task" [ 838.517431] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.527719] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882354, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.562817} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.542312] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] e9f2cd53-2c73-4c1f-858a-98a291da5547/e9f2cd53-2c73-4c1f-858a-98a291da5547.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 838.542312] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 838.542312] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882356, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.542312] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882353, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.542312] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 838.542312] env[62000]: value = "task-882357" [ 838.542312] env[62000]: _type = "Task" [ 838.542312] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.542312] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882355, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.543867] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8c716bfa-f9ba-4fe7-b9c3-36353fd47a09 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.559949] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882357, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.561416] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 838.561416] env[62000]: value = "task-882358" [ 838.561416] env[62000]: _type = "Task" [ 838.561416] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.570612] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882358, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.743317] env[62000]: DEBUG nova.network.neutron [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updated VIF entry in instance network info cache for port 7801caca-efe4-49ff-b583-7e91be8cf888. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 838.744102] env[62000]: DEBUG nova.network.neutron [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updating instance_info_cache with network_info: [{"id": "7801caca-efe4-49ff-b583-7e91be8cf888", "address": "fa:16:3e:c3:e3:98", "network": {"id": "51b6593d-5326-4a3d-9dfd-befc01ae8135", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-2013827846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed0fd1dd9fd74dbebdf9e919ae5da0d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7801caca-ef", "ovs_interfaceid": "7801caca-efe4-49ff-b583-7e91be8cf888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.802673] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.802983] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.803086] env[62000]: DEBUG nova.network.neutron [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.805585] env[62000]: DEBUG nova.scheduler.client.report [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 838.814386] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 838.838928] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 838.839205] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 838.839366] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 838.839550] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 838.839790] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 838.839843] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 838.840050] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 838.840220] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 838.840385] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 838.840547] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 838.840718] env[62000]: DEBUG nova.virt.hardware [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 838.841866] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753b28c5-274a-4c1f-b696-1c12ab5e6601 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.850629] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb41f00d-ed65-48ff-b892-5ca63ce74282 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.882234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.882234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.882234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.882234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.882234] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.884111] env[62000]: INFO nova.compute.manager [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Terminating instance [ 838.885982] env[62000]: DEBUG nova.compute.manager [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 838.886225] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 838.887050] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a753b0b-1609-4799-aa6f-047ec6a013dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.895783] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 838.896203] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3e498263-fc03-43b2-a88b-b6e93b9e12bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.903906] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 838.903906] env[62000]: value = "task-882359" [ 838.903906] env[62000]: _type = "Task" [ 838.903906] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.911763] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.964552] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882353, 'name': CreateVM_Task, 'duration_secs': 0.710103} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.964737] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 838.965460] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.965636] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.965957] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.966239] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-297a07c9-24ae-40c4-8c6f-f715b22f83f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.971414] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 838.971414] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52187696-4504-2c0e-c89e-38bc0261b8b8" [ 838.971414] env[62000]: _type = "Task" [ 838.971414] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.979301] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52187696-4504-2c0e-c89e-38bc0261b8b8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.993489] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882355, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.028385] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882356, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.055377] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882357, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.070436] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882358, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076603} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.070794] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.071628] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fead8227-cf66-47fb-be8b-f2986105efe8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.095330] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] e9f2cd53-2c73-4c1f-858a-98a291da5547/e9f2cd53-2c73-4c1f-858a-98a291da5547.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.095663] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff575c12-e3d7-4c41-a89d-33168da573c5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.119633] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 839.119633] env[62000]: value = "task-882360" [ 839.119633] env[62000]: _type = "Task" [ 839.119633] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.128369] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882360, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.249775] env[62000]: DEBUG oslo_concurrency.lockutils [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] Releasing lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.249775] env[62000]: DEBUG nova.compute.manager [req-8326154e-01bc-43c8-8f60-9a8f125d42c7 req-5a97056d-1270-4706-9a77-94532e611662 service nova] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Received event network-vif-deleted-ddd21181-32c0-4898-bd09-7689d6976198 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.311323] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.139s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.312023] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 839.314945] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.621s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.315188] env[62000]: DEBUG nova.objects.instance [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lazy-loading 'resources' on Instance uuid 96107e36-bf3f-4ef0-8d8b-5c9601f4f514 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.415705] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882359, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.488248] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52187696-4504-2c0e-c89e-38bc0261b8b8, 'name': SearchDatastore_Task, 'duration_secs': 0.033735} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.489085] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Successfully updated port: 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 839.495227] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.495227] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 839.495227] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.495227] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.495227] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 839.495227] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f056b928-6e78-4251-8642-d660845e38f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.508555] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882355, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.508555] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 839.508555] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 839.508819] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6922ff7-c8eb-4633-a720-0823fd875696 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.515375] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 839.515375] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52078e4e-312b-44f9-5c0b-69dcb21a6ed5" [ 839.515375] env[62000]: _type = "Task" [ 839.515375] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.532830] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52078e4e-312b-44f9-5c0b-69dcb21a6ed5, 'name': SearchDatastore_Task, 'duration_secs': 0.010572} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.533164] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882356, 'name': ReconfigVM_Task, 'duration_secs': 0.985234} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.534077] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.534752] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-db11021d-99fa-4b44-9bc1-2aa7cc90fc3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.537418] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15e9f698-8d66-423b-b7ba-bf4be4ffbea9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.543671] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 839.543671] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5243e4b7-dd95-c94f-f7d2-cfff1db1add6" [ 839.543671] env[62000]: _type = "Task" [ 839.543671] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.549494] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 839.549494] env[62000]: value = "task-882361" [ 839.549494] env[62000]: _type = "Task" [ 839.549494] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.571879] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882357, 'name': ReconfigVM_Task, 'duration_secs': 0.981728} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.572214] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5243e4b7-dd95-c94f-f7d2-cfff1db1add6, 'name': SearchDatastore_Task, 'duration_secs': 0.018098} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.572433] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882361, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.572710] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.573507] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.573648] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4e3f804c-f4a2-44ab-bc84-3c97d5803fab/4e3f804c-f4a2-44ab-bc84-3c97d5803fab.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 839.574588] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f008342b-88e8-476f-a04f-d6b98802ccbd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.576389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50ac949d-7645-4813-aa16-4aca9231f298 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.586360] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 839.586360] env[62000]: value = "task-882362" [ 839.586360] env[62000]: _type = "Task" [ 839.586360] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.588020] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 839.588020] env[62000]: value = "task-882363" [ 839.588020] env[62000]: _type = "Task" [ 839.588020] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.601269] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882362, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.604496] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882363, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.634165] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882360, 'name': ReconfigVM_Task, 'duration_secs': 0.297348} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.634488] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Reconfigured VM instance instance-00000045 to attach disk [datastore2] e9f2cd53-2c73-4c1f-858a-98a291da5547/e9f2cd53-2c73-4c1f-858a-98a291da5547.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 839.635148] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a0471c1f-100a-4b22-9402-c794b78153a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.644316] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 839.644316] env[62000]: value = "task-882364" [ 839.644316] env[62000]: _type = "Task" [ 839.644316] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.658643] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882364, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.701493] env[62000]: INFO nova.network.neutron [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Port 515167d1-0a88-4083-b014-249c351061a8 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 839.702112] env[62000]: DEBUG nova.network.neutron [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [{"id": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "address": "fa:16:3e:a4:28:43", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.206", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7961bfcb-4d", "ovs_interfaceid": "7961bfcb-4d94-4279-bc3a-b46ca382b0cf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.763765] env[62000]: DEBUG nova.compute.manager [req-cb62bc1f-2a33-4dfc-8adb-e3acd8a8e849 req-ff1f3105-89db-4bc7-9590-8d925a53372a service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-deleted-515167d1-0a88-4083-b014-249c351061a8 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 839.818665] env[62000]: DEBUG nova.compute.utils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 839.827678] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 839.827842] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 839.903771] env[62000]: DEBUG nova.policy [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffd8327dd8c4844b776982e99de5e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14aec090f2a4abc8078df1fe4428bbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 839.918729] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882359, 'name': PowerOffVM_Task, 'duration_secs': 0.903423} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.919195] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 839.919509] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 839.919873] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bc22ac17-17e5-47d0-90fa-ebe55de6230c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.997153] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.998653] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.998969] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.008602] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882355, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.017832] env[62000]: DEBUG nova.compute.manager [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Received event network-vif-plugged-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.018031] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.018300] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.018578] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.018693] env[62000]: DEBUG nova.compute.manager [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] No waiting events found dispatching network-vif-plugged-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.018871] env[62000]: WARNING nova.compute.manager [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Received unexpected event network-vif-plugged-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 for instance with vm_state building and task_state spawning. [ 840.019105] env[62000]: DEBUG nova.compute.manager [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Received event network-changed-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 840.020033] env[62000]: DEBUG nova.compute.manager [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Refreshing instance network info cache due to event network-changed-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 840.020033] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Acquiring lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.052910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 840.053559] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 840.053984] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleting the datastore file [datastore1] 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 840.054664] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-26939990-e18d-4c59-9cb6-3b25b75feaa4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.075119] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882361, 'name': Rename_Task, 'duration_secs': 0.277106} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.075119] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.075119] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 840.075119] env[62000]: value = "task-882366" [ 840.075119] env[62000]: _type = "Task" [ 840.075119] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.075119] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-786b51ba-c4c5-4513-813d-3437f366216d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.092181] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882366, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.102505] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 840.102505] env[62000]: value = "task-882367" [ 840.102505] env[62000]: _type = "Task" [ 840.102505] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.124772] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882362, 'name': Rename_Task, 'duration_secs': 0.285411} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.125099] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882363, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.133143] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.133143] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d3800b71-441f-4d6a-a2ae-69a5775d6905 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.137831] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882367, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.146265] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 840.146265] env[62000]: value = "task-882368" [ 840.146265] env[62000]: _type = "Task" [ 840.146265] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.165470] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882364, 'name': Rename_Task, 'duration_secs': 0.244937} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.168503] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 840.169594] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882368, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.172926] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a44a0b87-1a63-45a5-8b20-065a7d6bc020 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.183615] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 840.183615] env[62000]: value = "task-882369" [ 840.183615] env[62000]: _type = "Task" [ 840.183615] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.195011] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882369, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.205865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-71192360-6ee6-4876-bf37-da987a09cbb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.267685] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d3dc016-36ea-458a-813b-811cd7d148e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.276263] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b63a50c8-c101-4b07-8c2b-07404c706415 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.311497] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc6b9ce7-f45f-44e3-9f6a-dfeb96f01df2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.321363] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ee013b6-3017-4425-b03e-3f68ace97883 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.328667] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 840.339399] env[62000]: DEBUG nova.compute.provider_tree [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 840.498252] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882355, 'name': PowerOnVM_Task, 'duration_secs': 1.973187} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.498502] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 840.498717] env[62000]: DEBUG nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 840.499506] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6777e4f-0f89-42e8-b708-a35828e41190 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.562591] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.567033] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Successfully created port: 01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 840.588437] env[62000]: DEBUG oslo_vmware.api [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882366, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312575} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.591120] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 840.591345] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 840.591541] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 840.591714] env[62000]: INFO nova.compute.manager [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Took 1.71 seconds to destroy the instance on the hypervisor. [ 840.592050] env[62000]: DEBUG oslo.service.loopingcall [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 840.592367] env[62000]: DEBUG nova.compute.manager [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 840.592508] env[62000]: DEBUG nova.network.neutron [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 840.606373] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882363, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.626869} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.606373] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 4e3f804c-f4a2-44ab-bc84-3c97d5803fab/4e3f804c-f4a2-44ab-bc84-3c97d5803fab.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 840.606373] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.606373] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cfc710a0-ead0-40e2-9495-ac773c9fc612 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.617964] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882367, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.619579] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 840.619579] env[62000]: value = "task-882370" [ 840.619579] env[62000]: _type = "Task" [ 840.619579] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.629402] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882370, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.666628] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882368, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.695130] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882369, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.712620] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5c57c935-8fe6-42b1-90bf-1a065043cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-71192360-6ee6-4876-bf37-da987a09cbb2-656e8902-9da0-40c9-889a-8d1aab0f9266" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.106s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.721667] env[62000]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 515167d1-0a88-4083-b014-249c351061a8 could not be found.", "detail": ""}} {{(pid=62000) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 840.721907] env[62000]: DEBUG nova.network.neutron [-] Unable to show port 515167d1-0a88-4083-b014-249c351061a8 as it no longer exists. {{(pid=62000) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 840.845961] env[62000]: DEBUG nova.scheduler.client.report [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 840.957978] env[62000]: DEBUG nova.network.neutron [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating instance_info_cache with network_info: [{"id": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "address": "fa:16:3e:61:21:0a", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b065f0f-9b", "ovs_interfaceid": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.013516] env[62000]: INFO nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] bringing vm to original state: 'stopped' [ 841.119537] env[62000]: DEBUG oslo_vmware.api [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882367, 'name': PowerOnVM_Task, 'duration_secs': 0.895172} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.119537] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.119537] env[62000]: INFO nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Took 12.54 seconds to spawn the instance on the hypervisor. [ 841.119813] env[62000]: DEBUG nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.120465] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21785c2a-b941-4c6f-9cce-9be66f2ec825 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.135101] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882370, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.097167} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.135606] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.136426] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2aea030-4cd4-435a-a25b-b6383146b2a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.163538] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Reconfiguring VM instance instance-00000046 to attach disk [datastore2] 4e3f804c-f4a2-44ab-bc84-3c97d5803fab/4e3f804c-f4a2-44ab-bc84-3c97d5803fab.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.166818] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b78415f-b703-4246-a4c3-aef64c9e865e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.194264] env[62000]: DEBUG oslo_vmware.api [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882368, 'name': PowerOnVM_Task, 'duration_secs': 0.595659} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.198047] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.198047] env[62000]: INFO nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Took 15.24 seconds to spawn the instance on the hypervisor. [ 841.198047] env[62000]: DEBUG nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.198047] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 841.198047] env[62000]: value = "task-882371" [ 841.198047] env[62000]: _type = "Task" [ 841.198047] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.198047] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08e7ffa8-a441-4f59-9655-c09eb19ed307 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.209206] env[62000]: DEBUG oslo_vmware.api [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882369, 'name': PowerOnVM_Task, 'duration_secs': 0.558601} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.209206] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 841.209206] env[62000]: INFO nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Took 10.02 seconds to spawn the instance on the hypervisor. [ 841.209206] env[62000]: DEBUG nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 841.209730] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88bcb4ad-09b8-4101-bee3-3380901b15d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.217090] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882371, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.352823] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 841.355472] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.040s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.357723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.580s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.357977] env[62000]: DEBUG nova.objects.instance [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'resources' on Instance uuid ae1e8ebc-ddb8-4e95-847e-b9684e9161d4 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 841.394840] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 841.396746] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 841.397034] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 841.397238] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 841.397414] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 841.397591] env[62000]: DEBUG nova.virt.hardware [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 841.398531] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf3b5465-6e10-4b5d-9345-1b4f67479aaa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.402521] env[62000]: INFO nova.scheduler.client.report [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleted allocations for instance 96107e36-bf3f-4ef0-8d8b-5c9601f4f514 [ 841.412810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88ac90b5-8fcf-4a99-8ee4-5df7be9211e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.460771] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.461120] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Instance network_info: |[{"id": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "address": "fa:16:3e:61:21:0a", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b065f0f-9b", "ovs_interfaceid": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 841.461444] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Acquired lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.461661] env[62000]: DEBUG nova.network.neutron [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Refreshing network info cache for port 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 841.462887] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:61:21:0a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b065f0f-9bdd-4fbe-9992-aeebe892c6d0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 841.470617] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating folder: Project (8ce3270692934d2c9c1330a45c0e059e). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.471615] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f1a1f91a-784a-4139-843b-b81a12f6adee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.487841] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created folder: Project (8ce3270692934d2c9c1330a45c0e059e) in parent group-v201431. [ 841.488147] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating folder: Instances. Parent ref: group-v201572. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 841.488424] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bff83c68-4559-4c61-a62c-df49dcc3a022 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.500691] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created folder: Instances in parent group-v201572. [ 841.500948] env[62000]: DEBUG oslo.service.loopingcall [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.501163] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 841.501373] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-22e92baf-8544-4f67-9783-48c9b65ca6b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.522513] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 841.522513] env[62000]: value = "task-882374" [ 841.522513] env[62000]: _type = "Task" [ 841.522513] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.531389] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882374, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.645947] env[62000]: INFO nova.compute.manager [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Took 23.82 seconds to build instance. [ 841.711781] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882371, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.733246] env[62000]: INFO nova.compute.manager [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Took 24.64 seconds to build instance. [ 841.737534] env[62000]: INFO nova.compute.manager [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Took 21.21 seconds to build instance. [ 841.911879] env[62000]: DEBUG oslo_concurrency.lockutils [None req-73dc4846-8fad-4b05-9762-8b63af56ff4f tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "96107e36-bf3f-4ef0-8d8b-5c9601f4f514" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.670s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.920520] env[62000]: DEBUG nova.network.neutron [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.023024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.023024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.023024] env[62000]: DEBUG nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.023024] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43b4f54b-2668-4e20-8f4d-437a605ff618 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.041662] env[62000]: DEBUG nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 842.044512] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 842.044875] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882374, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.045189] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04accdd9-6db3-49bd-bd6c-fde83b65a721 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.059037] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 842.059037] env[62000]: value = "task-882375" [ 842.059037] env[62000]: _type = "Task" [ 842.059037] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.075016] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882375, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.149690] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cc49ad2f-aa7c-43a9-b945-5cdba3401bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.341s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.180545] env[62000]: DEBUG nova.compute.manager [req-8006d602-18dc-411b-b99d-076529b079a8 req-91e7dca3-7f18-4f07-92e6-f192a60c55cb service nova] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Received event network-vif-deleted-7961bfcb-4d94-4279-bc3a-b46ca382b0cf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.214603] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882371, 'name': ReconfigVM_Task, 'duration_secs': 0.753107} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.217276] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Reconfigured VM instance instance-00000046 to attach disk [datastore2] 4e3f804c-f4a2-44ab-bc84-3c97d5803fab/4e3f804c-f4a2-44ab-bc84-3c97d5803fab.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.218977] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e3fd4778-e530-449c-ac40-21b6dc32e3f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.228124] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 842.228124] env[62000]: value = "task-882376" [ 842.228124] env[62000]: _type = "Task" [ 842.228124] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.237987] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e61d34ed-6c54-45c1-a2fb-834794ee1721 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.157s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.244187] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3376e217-f6c1-4353-866d-b25f77ffad9f tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.721s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.244617] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882376, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.249610] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0aa7748-c8f2-4b38-846d-08a993a2fba8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.264866] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b9e308c-76ba-41aa-a9b8-42df1006c0e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.298363] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06fa3677-0f2c-4003-a83e-1a614b2cd256 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.303993] env[62000]: DEBUG nova.network.neutron [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updated VIF entry in instance network info cache for port 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 842.304408] env[62000]: DEBUG nova.network.neutron [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating instance_info_cache with network_info: [{"id": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "address": "fa:16:3e:61:21:0a", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b065f0f-9b", "ovs_interfaceid": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.309645] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b6d613-7ac8-4fde-96ae-d603a1805835 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.326905] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.425835] env[62000]: INFO nova.compute.manager [-] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Took 1.83 seconds to deallocate network for instance. [ 842.498880] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Successfully updated port: 01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.533863] env[62000]: DEBUG nova.compute.manager [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Received event network-vif-plugged-01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 842.534480] env[62000]: DEBUG oslo_concurrency.lockutils [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] Acquiring lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.534595] env[62000]: DEBUG oslo_concurrency.lockutils [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.534785] env[62000]: DEBUG oslo_concurrency.lockutils [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.535094] env[62000]: DEBUG nova.compute.manager [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] No waiting events found dispatching network-vif-plugged-01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.535330] env[62000]: WARNING nova.compute.manager [req-ae7b3495-134b-4505-b0d3-62eda537b8f6 req-7d134fda-778b-4450-97cd-2d142bc65f17 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Received unexpected event network-vif-plugged-01b29391-bf14-4d7b-b13c-8a45618192a2 for instance with vm_state building and task_state spawning. [ 842.543212] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882374, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.572311] env[62000]: DEBUG oslo_vmware.api [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882375, 'name': PowerOffVM_Task, 'duration_secs': 0.208402} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.573133] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 842.573328] env[62000]: DEBUG nova.compute.manager [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.574118] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b150060-ec17-4947-a89d-6bedd73ce144 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.740701] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882376, 'name': Rename_Task, 'duration_secs': 0.141371} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.741206] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 842.741538] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2f99f201-6828-45a2-9e61-d5c109493081 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.752109] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 842.752109] env[62000]: value = "task-882377" [ 842.752109] env[62000]: _type = "Task" [ 842.752109] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.759547] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882377, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.814198] env[62000]: DEBUG oslo_concurrency.lockutils [req-253e7a54-c7e4-474d-bf17-d62b9f63ddd4 req-f57e0244-f715-4833-a01f-aee7f363caef service nova] Releasing lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.854987] env[62000]: ERROR nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [req-e697b629-ad51-4854-bef5-e886e1a4d555] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-e697b629-ad51-4854-bef5-e886e1a4d555"}]} [ 842.879817] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 842.902604] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 842.902908] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.916258] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 842.934133] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.939362] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 843.001908] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.002092] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.002348] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.038970] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882374, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.060091] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.060433] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.060684] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.060902] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 843.061133] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.064462] env[62000]: INFO nova.compute.manager [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Terminating instance [ 843.066533] env[62000]: DEBUG nova.compute.manager [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 843.066743] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 843.067602] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8638080b-f24c-40af-a430-075b0efabfeb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.077908] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 843.080713] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-79d9afbb-1a41-4b1f-a9dd-659027e8a46f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.088768] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 843.088768] env[62000]: value = "task-882379" [ 843.088768] env[62000]: _type = "Task" [ 843.088768] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.089240] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.069s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.104069] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882379, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.262171] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882377, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.264038] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667f631c-02df-406a-95f2-219608783a77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.271899] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2217a92-81c8-4242-975b-64348fde5078 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.309544] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0699f71e-b529-4a85-8ab5-9f8a3233847b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.318609] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4a61d89-11eb-48d4-9d0c-8979c7e0b4bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.333690] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.537767] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882374, 'name': CreateVM_Task, 'duration_secs': 1.800553} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.537954] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 843.538676] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.538849] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.539200] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.539468] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5fa0252-4fe9-4ce7-bc0a-67fe6eeb5d5b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.544891] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 843.544891] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520e5344-7e97-3569-61bd-1474bfb52200" [ 843.544891] env[62000]: _type = "Task" [ 843.544891] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.554742] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520e5344-7e97-3569-61bd-1474bfb52200, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.561094] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.604213] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882379, 'name': PowerOffVM_Task, 'duration_secs': 0.206432} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.604705] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.604978] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 843.605168] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 843.605412] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-669f3512-355c-4d8b-aac0-a319c69fb50e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.753709] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 843.753946] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 843.754175] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Deleting the datastore file [datastore2] e9f2cd53-2c73-4c1f-858a-98a291da5547 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 843.757508] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27a88ae1-ddfb-4de3-ae35-337aebd49753 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.764560] env[62000]: DEBUG oslo_vmware.api [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882377, 'name': PowerOnVM_Task, 'duration_secs': 0.561582} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.764800] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 843.765017] env[62000]: INFO nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Took 10.03 seconds to spawn the instance on the hypervisor. [ 843.765209] env[62000]: DEBUG nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.766061] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfe184fb-4cda-4523-b5c9-4244cdc14a84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.769832] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for the task: (returnval){ [ 843.769832] env[62000]: value = "task-882381" [ 843.769832] env[62000]: _type = "Task" [ 843.769832] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.786553] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882381, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.793199] env[62000]: DEBUG nova.network.neutron [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Updating instance_info_cache with network_info: [{"id": "01b29391-bf14-4d7b-b13c-8a45618192a2", "address": "fa:16:3e:79:4d:92", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b29391-bf", "ovs_interfaceid": "01b29391-bf14-4d7b-b13c-8a45618192a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.856134] env[62000]: ERROR nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [req-6edf1f79-9dfe-4ff3-ac4d-f10b873fba64] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-6edf1f79-9dfe-4ff3-ac4d-f10b873fba64"}]} [ 843.871047] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 843.885035] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 843.885307] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.898998] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 843.916541] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 844.060009] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520e5344-7e97-3569-61bd-1474bfb52200, 'name': SearchDatastore_Task, 'duration_secs': 0.012365} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.060425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.060635] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 844.060870] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.061058] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.061278] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 844.064246] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25254def-2784-40d0-8786-58cebbd15900 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.076254] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.076254] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 844.076254] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bef7593d-b396-464a-8f92-b72a6097996d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.086565] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 844.086565] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d909f5-1930-6039-503a-e36ffa94b563" [ 844.086565] env[62000]: _type = "Task" [ 844.086565] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.099305] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d909f5-1930-6039-503a-e36ffa94b563, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.244051] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6efa64fc-f6a8-4967-b7b4-fe7ea312d189 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.253301] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a6aad28-e8a3-4d1a-9691-faebab51d5a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.289128] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdeeed29-554f-455b-be8b-5881ef33261e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.296851] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.297155] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance network_info: |[{"id": "01b29391-bf14-4d7b-b13c-8a45618192a2", "address": "fa:16:3e:79:4d:92", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b29391-bf", "ovs_interfaceid": "01b29391-bf14-4d7b-b13c-8a45618192a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.297856] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:79:4d:92', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '01b29391-bf14-4d7b-b13c-8a45618192a2', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.305287] env[62000]: DEBUG oslo.service.loopingcall [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.311883] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 844.312257] env[62000]: DEBUG oslo_vmware.api [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Task: {'id': task-882381, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.493676} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.312692] env[62000]: INFO nova.compute.manager [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Took 19.28 seconds to build instance. [ 844.313595] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9cf7afc-334e-4c80-9869-6c7c8b45a830 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.328487] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4fe04c-e596-4094-9e78-885ac256e812 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.332683] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 844.332889] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 844.333124] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 844.333328] env[62000]: INFO nova.compute.manager [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Took 1.27 seconds to destroy the instance on the hypervisor. [ 844.333566] env[62000]: DEBUG oslo.service.loopingcall [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.334018] env[62000]: DEBUG nova.compute.manager [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.334127] env[62000]: DEBUG nova.network.neutron [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 844.347908] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.352156] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.352156] env[62000]: value = "task-882382" [ 844.352156] env[62000]: _type = "Task" [ 844.352156] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.359287] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882382, 'name': CreateVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.444818] env[62000]: DEBUG nova.compute.manager [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 844.575559] env[62000]: DEBUG nova.compute.manager [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 844.575772] env[62000]: DEBUG nova.compute.manager [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing instance network info cache due to event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 844.575976] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.576148] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.576316] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 844.587341] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.587724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.587802] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.587961] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.588146] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.590095] env[62000]: INFO nova.compute.manager [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Terminating instance [ 844.594999] env[62000]: DEBUG nova.compute.manager [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 844.595207] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 844.596362] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcd9563-039b-4ad8-9800-11d207ac3b34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.605654] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d909f5-1930-6039-503a-e36ffa94b563, 'name': SearchDatastore_Task, 'duration_secs': 0.014361} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.608265] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 844.608487] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4adaadc-db8f-4759-b56c-45cbb5918567 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.610636] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b732d32c-f1d3-4996-8a19-fdcb24aa58ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.616314] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 844.616314] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fbcc7c-dde5-8fa6-d968-cb56f3477b69" [ 844.616314] env[62000]: _type = "Task" [ 844.616314] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.624416] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fbcc7c-dde5-8fa6-d968-cb56f3477b69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.725649] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 844.725881] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 844.726075] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore2] e2e3270d-086b-4441-a3d6-49b05a60b51f {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 844.726395] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-33108e0a-7f9d-487a-b872-3b9cd988ffc8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.733422] env[62000]: DEBUG oslo_vmware.api [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 844.733422] env[62000]: value = "task-882384" [ 844.733422] env[62000]: _type = "Task" [ 844.733422] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.742069] env[62000]: DEBUG oslo_vmware.api [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882384, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.834767] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3710ffd0-1fed-4c18-8b5a-8116fd73ef3d tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.817s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.866097] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882382, 'name': CreateVM_Task, 'duration_secs': 0.51581} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.866346] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 844.867080] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.867841] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.867841] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 844.868090] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e16be6b6-f1ae-4c42-a6b0-b665ee6d0610 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.874145] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 844.874145] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522abd27-66f0-7440-2023-f47ff4ede1c7" [ 844.874145] env[62000]: _type = "Task" [ 844.874145] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.884065] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522abd27-66f0-7440-2023-f47ff4ede1c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.885515] env[62000]: DEBUG nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 88 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 844.885749] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 88 to 89 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 844.885930] env[62000]: DEBUG nova.compute.provider_tree [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.964199] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.129854] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fbcc7c-dde5-8fa6-d968-cb56f3477b69, 'name': SearchDatastore_Task, 'duration_secs': 0.029761} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.130265] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.130418] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 3780b22d-c360-4433-9f6b-9d5d1f14b525/3780b22d-c360-4433-9f6b-9d5d1f14b525.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 845.130688] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d98025de-65ce-462c-be31-4501576b34b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.140197] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 845.140197] env[62000]: value = "task-882385" [ 845.140197] env[62000]: _type = "Task" [ 845.140197] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.148327] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882385, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.164829] env[62000]: DEBUG nova.network.neutron [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.243014] env[62000]: DEBUG oslo_vmware.api [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882384, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.389794} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.243352] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 845.243476] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 845.244065] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 845.244275] env[62000]: INFO nova.compute.manager [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Took 0.65 seconds to destroy the instance on the hypervisor. [ 845.244523] env[62000]: DEBUG oslo.service.loopingcall [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.245122] env[62000]: DEBUG nova.compute.manager [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 845.245122] env[62000]: DEBUG nova.network.neutron [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.385066] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522abd27-66f0-7440-2023-f47ff4ede1c7, 'name': SearchDatastore_Task, 'duration_secs': 0.013449} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.385389] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.386127] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.386127] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.386127] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.386396] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.386610] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f742e17c-cbbc-4ada-9485-7034622ab8fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.390338] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 4.033s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.392571] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.733s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.394150] env[62000]: INFO nova.compute.claims [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.400093] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.400093] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 845.400093] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0391c2cf-46b7-4172-bc96-bcd9ac4ad680 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.405212] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 845.405212] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522cc3f1-38a3-a23d-24f1-65bfa31d56d0" [ 845.405212] env[62000]: _type = "Task" [ 845.405212] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.414473] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522cc3f1-38a3-a23d-24f1-65bfa31d56d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.424542] env[62000]: INFO nova.scheduler.client.report [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted allocations for instance ae1e8ebc-ddb8-4e95-847e-b9684e9161d4 [ 845.590735] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updated VIF entry in instance network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 845.591179] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.601996] env[62000]: DEBUG nova.compute.manager [req-13aab558-1de8-4ddd-a952-3b9fbe3d92a5 req-27c7ef19-3271-4380-997d-40783025bfe9 service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Received event network-vif-deleted-0cb92206-16ee-49e1-8e70-fa3f39d678b4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 845.601996] env[62000]: INFO nova.compute.manager [req-13aab558-1de8-4ddd-a952-3b9fbe3d92a5 req-27c7ef19-3271-4380-997d-40783025bfe9 service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Neutron deleted interface 0cb92206-16ee-49e1-8e70-fa3f39d678b4; detaching it from the instance and deleting it from the info cache [ 845.601996] env[62000]: DEBUG nova.network.neutron [req-13aab558-1de8-4ddd-a952-3b9fbe3d92a5 req-27c7ef19-3271-4380-997d-40783025bfe9 service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.651581] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882385, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.667523] env[62000]: INFO nova.compute.manager [-] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Took 1.33 seconds to deallocate network for instance. [ 845.918215] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522cc3f1-38a3-a23d-24f1-65bfa31d56d0, 'name': SearchDatastore_Task, 'duration_secs': 0.013221} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.919291] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9171fae5-f352-4820-95b3-1e3b32a215a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.926719] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 845.926719] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524837a1-2095-c1c8-6f4a-516d0dda2aaf" [ 845.926719] env[62000]: _type = "Task" [ 845.926719] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.934000] env[62000]: DEBUG oslo_concurrency.lockutils [None req-661943cf-c691-4d1f-b3ec-294dce474d94 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "ae1e8ebc-ddb8-4e95-847e-b9684e9161d4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.914s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.940103] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524837a1-2095-c1c8-6f4a-516d0dda2aaf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.020268] env[62000]: DEBUG nova.network.neutron [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.093925] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.094608] env[62000]: DEBUG nova.compute.manager [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Received event network-changed-01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.094964] env[62000]: DEBUG nova.compute.manager [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Refreshing instance network info cache due to event network-changed-01b29391-bf14-4d7b-b13c-8a45618192a2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.095371] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Acquiring lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.095542] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Acquired lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.095849] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Refreshing network info cache for port 01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.105559] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-49bf9e5c-3805-4a03-9b11-7212877dd9c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.119056] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d62a2d3-d075-4bb8-b143-b31604e9fe4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.155632] env[62000]: DEBUG nova.compute.manager [req-13aab558-1de8-4ddd-a952-3b9fbe3d92a5 req-27c7ef19-3271-4380-997d-40783025bfe9 service nova] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Detach interface failed, port_id=0cb92206-16ee-49e1-8e70-fa3f39d678b4, reason: Instance e2e3270d-086b-4441-a3d6-49b05a60b51f could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 846.163089] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882385, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.925963} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.163089] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 3780b22d-c360-4433-9f6b-9d5d1f14b525/3780b22d-c360-4433-9f6b-9d5d1f14b525.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 846.163089] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.163089] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5b3edd86-f140-4f25-b369-2e112a079dc5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.171319] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 846.171319] env[62000]: value = "task-882386" [ 846.171319] env[62000]: _type = "Task" [ 846.171319] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.176209] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.182212] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882386, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.437969] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]524837a1-2095-c1c8-6f4a-516d0dda2aaf, 'name': SearchDatastore_Task, 'duration_secs': 0.055312} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.441092] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.441384] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] a432ca35-942d-434a-9cc9-e6e0302a44fd/a432ca35-942d-434a-9cc9-e6e0302a44fd.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 846.442380] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e73f8f3-f405-4f28-80f6-899d0cc97674 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.450233] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 846.450233] env[62000]: value = "task-882387" [ 846.450233] env[62000]: _type = "Task" [ 846.450233] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.461177] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882387, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.473430] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.473620] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.473751] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.473976] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.474208] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.478591] env[62000]: INFO nova.compute.manager [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Terminating instance [ 846.480507] env[62000]: DEBUG nova.compute.manager [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 846.480729] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 846.481531] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5fa727c-df1c-4c3c-8190-86a4d35d15ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.492413] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 846.495069] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-78b653d1-eedd-4953-871c-68cb6e1d3a85 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.503688] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 846.503688] env[62000]: value = "task-882388" [ 846.503688] env[62000]: _type = "Task" [ 846.503688] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.514844] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882388, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.525174] env[62000]: INFO nova.compute.manager [-] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Took 1.28 seconds to deallocate network for instance. [ 846.606262] env[62000]: DEBUG nova.compute.manager [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Received event network-vif-deleted-860a5cc9-5273-47c1-a9e7-b173fb60c2a1 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.606486] env[62000]: DEBUG nova.compute.manager [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Received event network-changed-7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 846.606655] env[62000]: DEBUG nova.compute.manager [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Refreshing instance network info cache due to event network-changed-7801caca-efe4-49ff-b583-7e91be8cf888. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 846.606867] env[62000]: DEBUG oslo_concurrency.lockutils [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] Acquiring lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.607020] env[62000]: DEBUG oslo_concurrency.lockutils [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] Acquired lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.607191] env[62000]: DEBUG nova.network.neutron [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Refreshing network info cache for port 7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 846.689991] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882386, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071962} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.689991] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.691042] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671419e9-4eb0-4b7d-b77c-63670e2db199 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.719352] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] 3780b22d-c360-4433-9f6b-9d5d1f14b525/3780b22d-c360-4433-9f6b-9d5d1f14b525.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.720844] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9ae569e-d534-4320-a93f-6db41342233d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.736665] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85eb4381-31b3-4ff2-a26d-390cea48e91b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.746657] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54785a18-fadd-435d-83fb-a64e6694d46e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.752231] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 846.752231] env[62000]: value = "task-882389" [ 846.752231] env[62000]: _type = "Task" [ 846.752231] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.783989] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d41c602-bfdb-4f6f-af40-8b90262626ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.790716] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882389, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.799522] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c74a796-897b-49cc-800b-1f952044c499 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.821464] env[62000]: DEBUG nova.compute.provider_tree [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 846.966839] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882387, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.014377] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882388, 'name': PowerOffVM_Task, 'duration_secs': 0.222965} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.014885] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 847.014885] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 847.015171] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b663f28b-1a92-4d0f-9455-ebd7823625ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.031274] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.154780] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Updated VIF entry in instance network info cache for port 01b29391-bf14-4d7b-b13c-8a45618192a2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.155195] env[62000]: DEBUG nova.network.neutron [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Updating instance_info_cache with network_info: [{"id": "01b29391-bf14-4d7b-b13c-8a45618192a2", "address": "fa:16:3e:79:4d:92", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap01b29391-bf", "ovs_interfaceid": "01b29391-bf14-4d7b-b13c-8a45618192a2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.190820] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 847.191117] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 847.191382] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleting the datastore file [datastore2] eb27703f-b657-423a-90a9-a7c024a2e473 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 847.191693] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6effccfe-e064-4887-944f-cac4aa3cad17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.201546] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for the task: (returnval){ [ 847.201546] env[62000]: value = "task-882391" [ 847.201546] env[62000]: _type = "Task" [ 847.201546] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.210525] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.262912] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882389, 'name': ReconfigVM_Task, 'duration_secs': 0.336678} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.263365] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfigured VM instance instance-00000047 to attach disk [datastore1] 3780b22d-c360-4433-9f6b-9d5d1f14b525/3780b22d-c360-4433-9f6b-9d5d1f14b525.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.264009] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bfe84004-6b2b-47f4-8c01-eb5c04443e48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.274195] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 847.274195] env[62000]: value = "task-882392" [ 847.274195] env[62000]: _type = "Task" [ 847.274195] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.284577] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882392, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.378154] env[62000]: DEBUG nova.scheduler.client.report [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 89 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 847.378467] env[62000]: DEBUG nova.compute.provider_tree [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 89 to 90 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 847.378662] env[62000]: DEBUG nova.compute.provider_tree [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.412392] env[62000]: DEBUG nova.network.neutron [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updated VIF entry in instance network info cache for port 7801caca-efe4-49ff-b583-7e91be8cf888. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 847.412808] env[62000]: DEBUG nova.network.neutron [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updating instance_info_cache with network_info: [{"id": "7801caca-efe4-49ff-b583-7e91be8cf888", "address": "fa:16:3e:c3:e3:98", "network": {"id": "51b6593d-5326-4a3d-9dfd-befc01ae8135", "bridge": "br-int", "label": "tempest-ServersTestFqdnHostnames-2013827846-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.202", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "ed0fd1dd9fd74dbebdf9e919ae5da0d0", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "39a4aca0-934b-4a91-8779-6a4360c3f967", "external-id": "nsx-vlan-transportzone-454", "segmentation_id": 454, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7801caca-ef", "ovs_interfaceid": "7801caca-efe4-49ff-b583-7e91be8cf888", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.466848] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882387, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.577793} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.467244] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] a432ca35-942d-434a-9cc9-e6e0302a44fd/a432ca35-942d-434a-9cc9-e6e0302a44fd.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 847.467517] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.467818] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7f108340-e1a4-4eb9-8f51-c00f81c14a74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.475889] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 847.475889] env[62000]: value = "task-882393" [ 847.475889] env[62000]: _type = "Task" [ 847.475889] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.484862] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882393, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.658459] env[62000]: DEBUG oslo_concurrency.lockutils [req-e4a6b2e8-dedb-437c-a694-d251df37b541 req-f96c6475-e987-4ce6-90e7-9ccb7cc14fe0 service nova] Releasing lock "refresh_cache-a432ca35-942d-434a-9cc9-e6e0302a44fd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.712661] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.784613] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882392, 'name': Rename_Task, 'duration_secs': 0.152528} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.784969] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 847.785340] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7ed16296-f6ff-4e1d-974b-ff525ab80abd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.793476] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 847.793476] env[62000]: value = "task-882394" [ 847.793476] env[62000]: _type = "Task" [ 847.793476] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.801918] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882394, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.887816] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.495s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.888396] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.891205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.949s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.891443] env[62000]: DEBUG nova.objects.instance [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lazy-loading 'resources' on Instance uuid 85f5f90a-9591-4393-9bef-2ad7e6a1a82f {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 847.916462] env[62000]: DEBUG oslo_concurrency.lockutils [req-539b0413-784e-4139-bab2-c6d94e20e770 req-bad89108-0356-4f9f-8c9d-8df9951c8ed5 service nova] Releasing lock "refresh_cache-4e3f804c-f4a2-44ab-bc84-3c97d5803fab" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.985573] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882393, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067585} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.985927] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 847.986767] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-123b0dfa-082b-4c6f-8d44-04fcd3ce83ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.008865] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] a432ca35-942d-434a-9cc9-e6e0302a44fd/a432ca35-942d-434a-9cc9-e6e0302a44fd.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.009174] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9cfa6c5-b1ec-469b-88ec-eeec17739450 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.029609] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 848.029609] env[62000]: value = "task-882395" [ 848.029609] env[62000]: _type = "Task" [ 848.029609] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.037420] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882395, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.211960] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.303822] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882394, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.394662] env[62000]: DEBUG nova.compute.utils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 848.399949] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 848.400258] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 848.475042] env[62000]: DEBUG nova.policy [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.545736] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882395, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.715738] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.720708] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82ca5d37-abbc-44bb-9c58-4bc06c016bb0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.727854] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bb160a8-0d31-4c71-9f1c-0717363df52b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.768547] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce22fcb9-c560-4caa-8ce3-4adfd4a8e7b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.776652] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc534257-95cc-4d19-be49-a5e919a5963b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.792506] env[62000]: DEBUG nova.compute.provider_tree [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.806830] env[62000]: DEBUG oslo_vmware.api [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882394, 'name': PowerOnVM_Task, 'duration_secs': 0.663572} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.807219] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 848.807516] env[62000]: INFO nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Took 9.99 seconds to spawn the instance on the hypervisor. [ 848.807784] env[62000]: DEBUG nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 848.809022] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c20395c-42b9-491c-a19f-2e60c40879a1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.897789] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 849.046576] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882395, 'name': ReconfigVM_Task, 'duration_secs': 0.562401} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.046962] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Reconfigured VM instance instance-00000048 to attach disk [datastore1] a432ca35-942d-434a-9cc9-e6e0302a44fd/a432ca35-942d-434a-9cc9-e6e0302a44fd.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 849.047860] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39e991a8-d644-4580-8799-19b8c7c0fc62 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.056603] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 849.056603] env[62000]: value = "task-882396" [ 849.056603] env[62000]: _type = "Task" [ 849.056603] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.057471] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Successfully created port: d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.069231] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882396, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.214189] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.296264] env[62000]: DEBUG nova.scheduler.client.report [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.338115] env[62000]: INFO nova.compute.manager [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Took 23.64 seconds to build instance. [ 849.570546] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882396, 'name': Rename_Task, 'duration_secs': 0.152364} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.570852] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 849.571290] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b3e63527-9e8b-4663-8692-e88fdd9e3084 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.578106] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 849.578106] env[62000]: value = "task-882397" [ 849.578106] env[62000]: _type = "Task" [ 849.578106] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.586589] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882397, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.715803] env[62000]: DEBUG oslo_vmware.api [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Task: {'id': task-882391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.245309} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.716135] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 849.716359] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 849.716569] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 849.716768] env[62000]: INFO nova.compute.manager [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Took 3.24 seconds to destroy the instance on the hypervisor. [ 849.717033] env[62000]: DEBUG oslo.service.loopingcall [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.717246] env[62000]: DEBUG nova.compute.manager [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 849.717354] env[62000]: DEBUG nova.network.neutron [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.801560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.803962] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.074s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.804258] env[62000]: DEBUG nova.objects.instance [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'resources' on Instance uuid 48d29490-c174-4702-a449-3bf10bc9207c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 849.825053] env[62000]: INFO nova.scheduler.client.report [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Deleted allocations for instance 85f5f90a-9591-4393-9bef-2ad7e6a1a82f [ 849.840560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0df4ed4c-9671-4313-9eaf-c2ea013d92cc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.162s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.906884] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.935219] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.935529] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.935705] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.935924] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.936171] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.936410] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.936658] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.936830] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.937051] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.937257] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.937441] env[62000]: DEBUG nova.virt.hardware [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.938374] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71c74d0f-0f61-499a-aef7-f45bc1d3ead0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.951213] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6b76a5-8d99-497c-87a7-7544ad3edcff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.003972] env[62000]: DEBUG nova.compute.manager [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Received event network-changed-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.004198] env[62000]: DEBUG nova.compute.manager [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Refreshing instance network info cache due to event network-changed-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 850.004447] env[62000]: DEBUG oslo_concurrency.lockutils [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] Acquiring lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.004857] env[62000]: DEBUG oslo_concurrency.lockutils [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] Acquired lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.005064] env[62000]: DEBUG nova.network.neutron [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Refreshing network info cache for port 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 850.090020] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882397, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.334404] env[62000]: DEBUG oslo_concurrency.lockutils [None req-798a769a-c4b5-4f8b-b40b-6a58acd3e985 tempest-ServersNegativeTestMultiTenantJSON-167501248 tempest-ServersNegativeTestMultiTenantJSON-167501248-project-member] Lock "85f5f90a-9591-4393-9bef-2ad7e6a1a82f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.624s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.597898] env[62000]: DEBUG oslo_vmware.api [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882397, 'name': PowerOnVM_Task, 'duration_secs': 0.567301} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.597898] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 850.598524] env[62000]: INFO nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Took 9.25 seconds to spawn the instance on the hypervisor. [ 850.598524] env[62000]: DEBUG nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 850.599119] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5175ae13-a942-4bd4-96fa-0868e3f12930 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.646463] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468b78c2-925f-4393-9e6f-78030751ce72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.655268] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1f89cd8-143b-42e3-a4a0-d5ae4d98d8e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.691898] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436e4522-36d2-4f63-9b43-91df57814cbb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.700162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9582e3f7-2da4-46fc-ae7b-bd3c347c19e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.715767] env[62000]: DEBUG nova.compute.provider_tree [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.727965] env[62000]: DEBUG nova.compute.manager [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Received event network-vif-plugged-d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 850.728055] env[62000]: DEBUG oslo_concurrency.lockutils [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] Acquiring lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 850.728274] env[62000]: DEBUG oslo_concurrency.lockutils [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.728480] env[62000]: DEBUG oslo_concurrency.lockutils [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.728682] env[62000]: DEBUG nova.compute.manager [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] No waiting events found dispatching network-vif-plugged-d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 850.728866] env[62000]: WARNING nova.compute.manager [req-828b0a74-4a17-4d2a-b1b1-73fff6e066b0 req-dd030014-276a-4c6f-b80c-9533643c2ecf service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Received unexpected event network-vif-plugged-d42587d8-a09e-4869-b4d4-6f8265136d23 for instance with vm_state building and task_state spawning. [ 850.824995] env[62000]: DEBUG nova.network.neutron [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.873509] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Successfully updated port: d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.937156] env[62000]: DEBUG nova.network.neutron [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updated VIF entry in instance network info cache for port 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 850.937725] env[62000]: DEBUG nova.network.neutron [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating instance_info_cache with network_info: [{"id": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "address": "fa:16:3e:61:21:0a", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.154", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b065f0f-9b", "ovs_interfaceid": "7b065f0f-9bdd-4fbe-9992-aeebe892c6d0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.120147] env[62000]: INFO nova.compute.manager [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Took 22.75 seconds to build instance. [ 851.217384] env[62000]: DEBUG nova.scheduler.client.report [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.328716] env[62000]: INFO nova.compute.manager [-] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Took 1.61 seconds to deallocate network for instance. [ 851.377740] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.377900] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.378069] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 851.443442] env[62000]: DEBUG oslo_concurrency.lockutils [req-adc9843b-d0fd-4ebb-99f8-952d18bdff32 req-6f8170b5-d9d5-4dd9-8358-7c747ddb402b service nova] Releasing lock "refresh_cache-3780b22d-c360-4433-9f6b-9d5d1f14b525" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.479244] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99dc6700-794b-4983-ae1b-12d6b19cfb9b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.488816] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Suspending the VM {{(pid=62000) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 851.489098] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-a0aaf1e2-db30-42d5-a9ee-834ed6fadc99 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.495299] env[62000]: DEBUG oslo_vmware.api [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 851.495299] env[62000]: value = "task-882398" [ 851.495299] env[62000]: _type = "Task" [ 851.495299] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.503271] env[62000]: DEBUG oslo_vmware.api [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882398, 'name': SuspendVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.623684] env[62000]: DEBUG oslo_concurrency.lockutils [None req-636bc0e3-ec8b-447c-9105-1aa5cef457c8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.258s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.723366] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.919s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.729235] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.904s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.730918] env[62000]: INFO nova.compute.claims [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.772178] env[62000]: INFO nova.scheduler.client.report [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance 48d29490-c174-4702-a449-3bf10bc9207c [ 851.835611] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.923191] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 852.005827] env[62000]: DEBUG oslo_vmware.api [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882398, 'name': SuspendVM_Task} progress is 70%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.036926] env[62000]: DEBUG nova.compute.manager [req-429975d4-337b-4a43-9510-bcc0f6fd6d71 req-71c11e59-6fe0-4240-80ff-6e1ffb80fc8e service nova] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Received event network-vif-deleted-d725993d-bf92-4c34-9f24-1c0ed6557610 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.153996] env[62000]: DEBUG nova.network.neutron [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Updating instance_info_cache with network_info: [{"id": "d42587d8-a09e-4869-b4d4-6f8265136d23", "address": "fa:16:3e:0e:8b:b9", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd42587d8-a0", "ovs_interfaceid": "d42587d8-a09e-4869-b4d4-6f8265136d23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.281636] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6124fe9b-0374-4c73-9ba0-c454b9b782ce tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "48d29490-c174-4702-a449-3bf10bc9207c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.412s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.506019] env[62000]: DEBUG oslo_vmware.api [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882398, 'name': SuspendVM_Task, 'duration_secs': 0.601612} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.506241] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Suspended the VM {{(pid=62000) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 852.506433] env[62000]: DEBUG nova.compute.manager [None req-048efc11-c720-479c-a445-01dcd44f238c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.507221] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3df4f5-21aa-49ee-b626-883127c89243 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.656602] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.657193] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Instance network_info: |[{"id": "d42587d8-a09e-4869-b4d4-6f8265136d23", "address": "fa:16:3e:0e:8b:b9", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd42587d8-a0", "ovs_interfaceid": "d42587d8-a09e-4869-b4d4-6f8265136d23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.657669] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:8b:b9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd42587d8-a09e-4869-b4d4-6f8265136d23', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.665158] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating folder: Project (04d07461c67e4868a33a345d2e08db82). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 852.665475] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-46974a58-a2a9-4693-a40c-2574f8a2b1f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.677971] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created folder: Project (04d07461c67e4868a33a345d2e08db82) in parent group-v201431. [ 852.678341] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating folder: Instances. Parent ref: group-v201576. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 852.678614] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-075e824c-a712-4a84-94cf-a32de19d80f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.688931] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created folder: Instances in parent group-v201576. [ 852.689334] env[62000]: DEBUG oslo.service.loopingcall [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.689572] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 852.689809] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-13d07390-ef61-4aeb-9a72-2ca0178cf235 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.711511] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.711511] env[62000]: value = "task-882401" [ 852.711511] env[62000]: _type = "Task" [ 852.711511] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.718749] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882401, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.769109] env[62000]: DEBUG nova.compute.manager [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Received event network-changed-d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 852.769109] env[62000]: DEBUG nova.compute.manager [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Refreshing instance network info cache due to event network-changed-d42587d8-a09e-4869-b4d4-6f8265136d23. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 852.769109] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] Acquiring lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.769109] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] Acquired lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.769109] env[62000]: DEBUG nova.network.neutron [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Refreshing network info cache for port d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 853.056846] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d7ae619-eda1-4fd8-92ee-30b49465585a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.065317] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af7c3eab-3319-47a8-a653-bc5dad78bf58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.101240] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1b644c-89fd-4e58-b5d2-4f9ad6d6693d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.107601] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9ed765e-1d69-44f2-bd8e-8107898dcc0a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.123772] env[62000]: DEBUG nova.compute.provider_tree [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 853.221882] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882401, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.626756] env[62000]: DEBUG nova.scheduler.client.report [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.681155] env[62000]: DEBUG nova.network.neutron [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Updated VIF entry in instance network info cache for port d42587d8-a09e-4869-b4d4-6f8265136d23. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 853.681550] env[62000]: DEBUG nova.network.neutron [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Updating instance_info_cache with network_info: [{"id": "d42587d8-a09e-4869-b4d4-6f8265136d23", "address": "fa:16:3e:0e:8b:b9", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd42587d8-a0", "ovs_interfaceid": "d42587d8-a09e-4869-b4d4-6f8265136d23", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.724142] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882401, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.060791] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.061748] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.132464] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.133062] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 854.136024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.344s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.136258] env[62000]: DEBUG nova.objects.instance [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lazy-loading 'resources' on Instance uuid ea8ae421-1d7f-4814-bc0f-90a3316ad028 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 854.190014] env[62000]: DEBUG oslo_concurrency.lockutils [req-5f660a6f-0c97-4b74-82c9-4b430583314d req-709154f2-eac2-4439-a7f4-9e50eaf6a8fb service nova] Releasing lock "refresh_cache-44a54bd7-a35c-49ed-85ed-346830cee6ad" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.224026] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882401, 'name': CreateVM_Task, 'duration_secs': 1.339229} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.224334] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 854.224890] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.225076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.225421] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.225677] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fd21901c-eed8-477b-99c8-d04f8ebc99e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.230970] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 854.230970] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a57a25-e675-bc90-f66c-0814e3ddafa1" [ 854.230970] env[62000]: _type = "Task" [ 854.230970] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.238850] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a57a25-e675-bc90-f66c-0814e3ddafa1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.563755] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 854.639478] env[62000]: DEBUG nova.compute.utils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 854.645606] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 854.645606] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 854.715992] env[62000]: DEBUG nova.policy [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f3c14a8d159a4578b645cb7ec67c8ec6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fcbc6bb628fb41248aee27ded8cb0d7f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 854.744208] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a57a25-e675-bc90-f66c-0814e3ddafa1, 'name': SearchDatastore_Task, 'duration_secs': 0.009395} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.744208] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.744329] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.744493] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.744636] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.744811] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.747017] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7679dc69-86cf-498e-acf0-723304c02704 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.756970] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.757179] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 854.757916] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c82ed5d-4d8f-445b-a52c-d75b51429798 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.769019] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 854.769019] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52071953-2826-87dc-94b0-bd1f31278d3f" [ 854.769019] env[62000]: _type = "Task" [ 854.769019] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.777021] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52071953-2826-87dc-94b0-bd1f31278d3f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.933901] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee8533f1-b47f-400a-af96-4078eebc3a05 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.941794] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ac831f-a1e8-4824-9b44-56a146c3cc83 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.976078] env[62000]: DEBUG nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.977068] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f562431-9ee6-4bca-afe3-b5b7b48da0dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.981030] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef384c2b-844d-475e-bb1a-de603abd9f4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.991972] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34b9a595-d458-4474-8832-aa1b439bfd47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.011691] env[62000]: DEBUG nova.compute.provider_tree [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 855.031234] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Successfully created port: 607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 855.090691] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.143772] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 855.282778] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52071953-2826-87dc-94b0-bd1f31278d3f, 'name': SearchDatastore_Task, 'duration_secs': 0.01108} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.283882] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47bbe4f7-395d-4d22-86d8-e615aba36c4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.292170] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 855.292170] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52144e02-e090-d511-283e-9b59357bd43e" [ 855.292170] env[62000]: _type = "Task" [ 855.292170] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.301989] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52144e02-e090-d511-283e-9b59357bd43e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.499925] env[62000]: INFO nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] instance snapshotting [ 855.500174] env[62000]: WARNING nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 855.502942] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e8742a-9fd3-43d2-b56d-befdf9f1bbee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.521752] env[62000]: DEBUG nova.scheduler.client.report [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 855.526309] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32c29dd-fa99-4679-96c6-19b2250761b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.803373] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52144e02-e090-d511-283e-9b59357bd43e, 'name': SearchDatastore_Task, 'duration_secs': 0.013354} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.803685] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.804010] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 44a54bd7-a35c-49ed-85ed-346830cee6ad/44a54bd7-a35c-49ed-85ed-346830cee6ad.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 855.804405] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-16a91d37-ab46-487d-8724-e69b4e2db822 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.811648] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 855.811648] env[62000]: value = "task-882402" [ 855.811648] env[62000]: _type = "Task" [ 855.811648] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.819041] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882402, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.030204] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.894s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.033098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.239s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.033098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.035293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 18.204s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.039200] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 856.039764] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-be724ed4-e8cb-405e-a8c0-3bdcee8a1e5c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.048703] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 856.048703] env[62000]: value = "task-882403" [ 856.048703] env[62000]: _type = "Task" [ 856.048703] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.064243] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882403, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.065460] env[62000]: INFO nova.scheduler.client.report [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Deleted allocations for instance f3a6527c-777e-4a07-9482-598de15d4eb3 [ 856.070940] env[62000]: INFO nova.scheduler.client.report [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Deleted allocations for instance ea8ae421-1d7f-4814-bc0f-90a3316ad028 [ 856.154400] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 856.183665] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 856.184025] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 856.184136] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 856.184292] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 856.184815] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 856.184815] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 856.184815] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 856.185051] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 856.185805] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 856.185805] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 856.185805] env[62000]: DEBUG nova.virt.hardware [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 856.187209] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7edca8a-64b5-4379-840c-e4eaa85e490d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.197159] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319d734a-b965-43fe-bdc1-fd846a81a15d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.321111] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882402, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.542801] env[62000]: INFO nova.compute.claims [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.559282] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882403, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.581330] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ca8a665a-486d-44ef-b9dd-2455310c1074 tempest-MigrationsAdminTest-1401037496 tempest-MigrationsAdminTest-1401037496-project-member] Lock "f3a6527c-777e-4a07-9482-598de15d4eb3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.816s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.582764] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a3f29c62-6857-4025-a52c-8deadd56604a tempest-AttachInterfacesUnderV243Test-145862007 tempest-AttachInterfacesUnderV243Test-145862007-project-member] Lock "ea8ae421-1d7f-4814-bc0f-90a3316ad028" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.734s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.822750] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882402, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.879288] env[62000]: DEBUG nova.compute.manager [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Received event network-vif-plugged-607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 856.879485] env[62000]: DEBUG oslo_concurrency.lockutils [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] Acquiring lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.879674] env[62000]: DEBUG oslo_concurrency.lockutils [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.881223] env[62000]: DEBUG oslo_concurrency.lockutils [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.881223] env[62000]: DEBUG nova.compute.manager [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] No waiting events found dispatching network-vif-plugged-607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.881223] env[62000]: WARNING nova.compute.manager [req-62e31afc-8ab6-4ca3-a84c-142e889ebc02 req-ac8017f6-d15d-459c-a292-5578380cff11 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Received unexpected event network-vif-plugged-607b5eaf-2995-418d-8242-05869638adf3 for instance with vm_state building and task_state spawning. [ 857.020638] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Successfully updated port: 607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 857.050019] env[62000]: INFO nova.compute.resource_tracker [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating resource usage from migration 430688a3-72d0-4b3b-a5ec-98b8352758e1 [ 857.062440] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882403, 'name': CreateSnapshot_Task, 'duration_secs': 0.642274} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.063040] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 857.063515] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a928ee53-14df-4f5b-ae5d-4415fcba5f27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.324753] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882402, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.491307} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.327425] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 44a54bd7-a35c-49ed-85ed-346830cee6ad/44a54bd7-a35c-49ed-85ed-346830cee6ad.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 857.327664] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 857.328135] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cfeb689-976e-49fd-9c97-aeef08d12fc5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.334712] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 857.334712] env[62000]: value = "task-882404" [ 857.334712] env[62000]: _type = "Task" [ 857.334712] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.344782] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882404, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.370720] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb104a9-fc8b-438a-8d07-cc9ecd39f355 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.377958] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25733252-8432-4304-a241-936b6325513f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.413284] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b090a04d-5c74-461b-bc6c-7d3d88b64d9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.420520] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3955e91-bc43-4a16-aaf9-674363b09e84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.435540] env[62000]: DEBUG nova.compute.provider_tree [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 857.526252] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.526252] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquired lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 857.526252] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 857.589653] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 857.589928] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0b668b00-c353-4495-9aca-728e7995869e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.600021] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 857.600021] env[62000]: value = "task-882405" [ 857.600021] env[62000]: _type = "Task" [ 857.600021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.608902] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882405, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.845792] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882404, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070873} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.849059] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 857.849059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd18f0f7-c1fc-4213-b85b-41740e673490 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.871679] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 44a54bd7-a35c-49ed-85ed-346830cee6ad/44a54bd7-a35c-49ed-85ed-346830cee6ad.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 857.872210] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4747c9ec-a755-46da-bf5d-d04c0e8cc441 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.893998] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 857.893998] env[62000]: value = "task-882406" [ 857.893998] env[62000]: _type = "Task" [ 857.893998] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.905719] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882406, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.944063] env[62000]: DEBUG nova.scheduler.client.report [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 858.075068] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.114670] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882405, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.383198] env[62000]: DEBUG nova.network.neutron [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Updating instance_info_cache with network_info: [{"id": "607b5eaf-2995-418d-8242-05869638adf3", "address": "fa:16:3e:f5:50:b6", "network": {"id": "0bdfce42-2cd6-4423-a6d9-7da80330c568", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-403212104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcbc6bb628fb41248aee27ded8cb0d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap607b5eaf-29", "ovs_interfaceid": "607b5eaf-2995-418d-8242-05869638adf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 858.409099] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882406, 'name': ReconfigVM_Task, 'duration_secs': 0.47575} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.410937] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 44a54bd7-a35c-49ed-85ed-346830cee6ad/44a54bd7-a35c-49ed-85ed-346830cee6ad.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 858.410937] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-be4f6223-d0d0-489d-9d66-df505ffaa1ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.416731] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 858.416731] env[62000]: value = "task-882407" [ 858.416731] env[62000]: _type = "Task" [ 858.416731] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.426125] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882407, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.450974] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.416s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.452042] env[62000]: INFO nova.compute.manager [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Migrating [ 858.463753] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.530s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.464805] env[62000]: DEBUG nova.objects.instance [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'resources' on Instance uuid 71192360-6ee6-4876-bf37-da987a09cbb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 858.613918] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882405, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.886587] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Releasing lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.886937] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Instance network_info: |[{"id": "607b5eaf-2995-418d-8242-05869638adf3", "address": "fa:16:3e:f5:50:b6", "network": {"id": "0bdfce42-2cd6-4423-a6d9-7da80330c568", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-403212104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcbc6bb628fb41248aee27ded8cb0d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap607b5eaf-29", "ovs_interfaceid": "607b5eaf-2995-418d-8242-05869638adf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 858.887506] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f5:50:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '447ff42d-b33e-4b5d-8b7f-e8117ebbbc92', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '607b5eaf-2995-418d-8242-05869638adf3', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 858.895308] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Creating folder: Project (fcbc6bb628fb41248aee27ded8cb0d7f). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 858.895602] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-dce6c495-0f94-4c3a-bbc7-1ed2ebd959ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.905025] env[62000]: DEBUG nova.compute.manager [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Received event network-changed-607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 858.905228] env[62000]: DEBUG nova.compute.manager [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Refreshing instance network info cache due to event network-changed-607b5eaf-2995-418d-8242-05869638adf3. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 858.905444] env[62000]: DEBUG oslo_concurrency.lockutils [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] Acquiring lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.905619] env[62000]: DEBUG oslo_concurrency.lockutils [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] Acquired lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.905783] env[62000]: DEBUG nova.network.neutron [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Refreshing network info cache for port 607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 858.908444] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Created folder: Project (fcbc6bb628fb41248aee27ded8cb0d7f) in parent group-v201431. [ 858.908637] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Creating folder: Instances. Parent ref: group-v201581. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 858.909049] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a33c15db-dd3b-4b38-bf16-564565d31733 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.923817] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Created folder: Instances in parent group-v201581. [ 858.923817] env[62000]: DEBUG oslo.service.loopingcall [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 858.923817] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 858.923817] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7d1fbb37-9be1-42c4-bce7-ab06e78ac42f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.941094] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882407, 'name': Rename_Task, 'duration_secs': 0.182839} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.941757] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 858.942092] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-17875ff6-9ad1-4e81-970f-ceae67e068e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.945176] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 858.945176] env[62000]: value = "task-882410" [ 858.945176] env[62000]: _type = "Task" [ 858.945176] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.949566] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 858.949566] env[62000]: value = "task-882411" [ 858.949566] env[62000]: _type = "Task" [ 858.949566] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.957456] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882410, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.960616] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882411, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.980504] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.980504] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.980504] env[62000]: DEBUG nova.network.neutron [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 859.114534] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882405, 'name': CloneVM_Task, 'duration_secs': 1.205729} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.114877] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Created linked-clone VM from snapshot [ 859.115609] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5facead1-630d-4bf3-8fc6-7a95b87e41ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.124469] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Uploading image 2b9a76bd-b670-46f7-9dda-25750fa1d88d {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 859.161812] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 859.161812] env[62000]: value = "vm-201580" [ 859.161812] env[62000]: _type = "VirtualMachine" [ 859.161812] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 859.162468] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-30451a0d-19b2-46d6-8975-d6eadb315eb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.173554] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease: (returnval){ [ 859.173554] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d011b-8775-6026-b048-243d92585ea8" [ 859.173554] env[62000]: _type = "HttpNfcLease" [ 859.173554] env[62000]: } obtained for exporting VM: (result){ [ 859.173554] env[62000]: value = "vm-201580" [ 859.173554] env[62000]: _type = "VirtualMachine" [ 859.173554] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 859.173964] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the lease: (returnval){ [ 859.173964] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d011b-8775-6026-b048-243d92585ea8" [ 859.173964] env[62000]: _type = "HttpNfcLease" [ 859.173964] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 859.185526] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.185526] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d011b-8775-6026-b048-243d92585ea8" [ 859.185526] env[62000]: _type = "HttpNfcLease" [ 859.185526] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 859.284671] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b9e7a69-f668-4097-9914-da65527d99de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.293216] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cce3e69-0ee8-49f9-bada-953bcca8373e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.324763] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9663d435-bbb8-4a39-910f-c272651d0e2e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.333753] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cb7bb78-0191-4449-a705-488dc978fdf0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.351847] env[62000]: DEBUG nova.compute.provider_tree [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.459344] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882410, 'name': CreateVM_Task, 'duration_secs': 0.334528} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.460127] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 859.460809] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.461006] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.461365] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 859.464439] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3cfecca-1040-4a73-abff-753aab78e3d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.465959] env[62000]: DEBUG oslo_vmware.api [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882411, 'name': PowerOnVM_Task, 'duration_secs': 0.494648} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.466254] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 859.466469] env[62000]: INFO nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Took 9.56 seconds to spawn the instance on the hypervisor. [ 859.466648] env[62000]: DEBUG nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.467712] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62e62fb-d1da-4919-a465-8dad3b0aadd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.471170] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 859.471170] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ac1dfd-b3ca-4414-e347-fe86e2438118" [ 859.471170] env[62000]: _type = "Task" [ 859.471170] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.485890] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ac1dfd-b3ca-4414-e347-fe86e2438118, 'name': SearchDatastore_Task, 'duration_secs': 0.008606} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.486198] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.486430] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 859.486657] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 859.486805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 859.486987] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 859.487545] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-26c8d6ed-14f3-4754-b321-50a4703b54ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.495397] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 859.495608] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 859.496322] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d50496e-161e-4643-9223-ae41ef266770 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.501484] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 859.501484] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ab8696-ab04-dea7-622a-ade6543ab480" [ 859.501484] env[62000]: _type = "Task" [ 859.501484] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.511032] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ab8696-ab04-dea7-622a-ade6543ab480, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.668647] env[62000]: DEBUG nova.network.neutron [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Updated VIF entry in instance network info cache for port 607b5eaf-2995-418d-8242-05869638adf3. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 859.669096] env[62000]: DEBUG nova.network.neutron [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Updating instance_info_cache with network_info: [{"id": "607b5eaf-2995-418d-8242-05869638adf3", "address": "fa:16:3e:f5:50:b6", "network": {"id": "0bdfce42-2cd6-4423-a6d9-7da80330c568", "bridge": "br-int", "label": "tempest-ServerPasswordTestJSON-403212104-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "fcbc6bb628fb41248aee27ded8cb0d7f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "447ff42d-b33e-4b5d-8b7f-e8117ebbbc92", "external-id": "nsx-vlan-transportzone-836", "segmentation_id": 836, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap607b5eaf-29", "ovs_interfaceid": "607b5eaf-2995-418d-8242-05869638adf3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.683838] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 859.683838] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d011b-8775-6026-b048-243d92585ea8" [ 859.683838] env[62000]: _type = "HttpNfcLease" [ 859.683838] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 859.684132] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 859.684132] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527d011b-8775-6026-b048-243d92585ea8" [ 859.684132] env[62000]: _type = "HttpNfcLease" [ 859.684132] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 859.684832] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d2e340b-ae66-4d66-b802-a708fb198b11 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.692540] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 859.692718] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 859.821553] env[62000]: DEBUG nova.network.neutron [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.859737] env[62000]: DEBUG nova.scheduler.client.report [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.881140] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-6f532a01-4584-4a68-9023-d2466f8f49e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.990233] env[62000]: INFO nova.compute.manager [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Took 29.37 seconds to build instance. [ 860.013132] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ab8696-ab04-dea7-622a-ade6543ab480, 'name': SearchDatastore_Task, 'duration_secs': 0.008414} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.014509] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9905f6d-5371-4d27-abb3-eb964cb0295b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.020480] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 860.020480] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5259e49e-f478-7026-2847-465fc5c59ba2" [ 860.020480] env[62000]: _type = "Task" [ 860.020480] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.029709] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5259e49e-f478-7026-2847-465fc5c59ba2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.171534] env[62000]: DEBUG oslo_concurrency.lockutils [req-43381113-1bfd-4210-9381-44edb7df1b3f req-1e62dc68-f1b7-4beb-8f64-f9d1b7167200 service nova] Releasing lock "refresh_cache-cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.324358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.364719] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.367854] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.763s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.368201] env[62000]: DEBUG nova.objects.instance [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 860.386266] env[62000]: INFO nova.scheduler.client.report [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted allocations for instance 71192360-6ee6-4876-bf37-da987a09cbb2 [ 860.494568] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cd71e124-6df3-4c17-821b-459d1ca3b7f1 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.894s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.534317] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5259e49e-f478-7026-2847-465fc5c59ba2, 'name': SearchDatastore_Task, 'duration_secs': 0.019789} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.534770] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 860.535142] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] cc7e8af9-16ef-41bf-a91e-edb6c166a3a5/cc7e8af9-16ef-41bf-a91e-edb6c166a3a5.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 860.535473] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5274faf8-b03a-4ef4-8e0b-fd9a30ac372d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.543755] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 860.543755] env[62000]: value = "task-882413" [ 860.543755] env[62000]: _type = "Task" [ 860.543755] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.555565] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882413, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.896687] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b8b77ca8-f205-4ed0-8de6-6173c2bf1dbf tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "71192360-6ee6-4876-bf37-da987a09cbb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.016s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.059128] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882413, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485696} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.059585] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] cc7e8af9-16ef-41bf-a91e-edb6c166a3a5/cc7e8af9-16ef-41bf-a91e-edb6c166a3a5.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 861.059920] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 861.060337] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f81fd753-6212-41e4-9bbe-f9e2ae016bb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.068575] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 861.068575] env[62000]: value = "task-882414" [ 861.068575] env[62000]: _type = "Task" [ 861.068575] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.080876] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882414, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.294525] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 861.294900] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.378981] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6bc13c42-a36f-4d3e-a605-ecc1171960ef tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.380079] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 16.416s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 861.578635] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882414, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.140702} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.579075] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 861.579746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e621afad-6986-47bb-be08-21b73b21195c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.602250] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] cc7e8af9-16ef-41bf-a91e-edb6c166a3a5/cc7e8af9-16ef-41bf-a91e-edb6c166a3a5.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 861.602587] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c593c86-74d2-4d40-bdc3-2c500e509467 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.624255] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 861.624255] env[62000]: value = "task-882415" [ 861.624255] env[62000]: _type = "Task" [ 861.624255] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.632469] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882415, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.797129] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 861.845931] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ea6cf8b-3882-4e1f-857c-e19d202946ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.865681] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 861.887482] env[62000]: INFO nova.compute.claims [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.135400] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882415, 'name': ReconfigVM_Task, 'duration_secs': 0.320152} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.135737] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Reconfigured VM instance instance-0000004a to attach disk [datastore1] cc7e8af9-16ef-41bf-a91e-edb6c166a3a5/cc7e8af9-16ef-41bf-a91e-edb6c166a3a5.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 862.136406] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-52fbbb70-48d6-4f60-98fa-92a09ae1f959 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.143343] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 862.143343] env[62000]: value = "task-882416" [ 862.143343] env[62000]: _type = "Task" [ 862.143343] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.151442] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882416, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.323693] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.377398] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 862.377844] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e3510709-99e7-4f98-b498-6527bcd0de0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.385585] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 862.385585] env[62000]: value = "task-882417" [ 862.385585] env[62000]: _type = "Task" [ 862.385585] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.392421] env[62000]: INFO nova.compute.resource_tracker [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating resource usage from migration 4f39862a-db3a-43d6-9aa3-457c8f47a8c9 [ 862.399185] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 862.400027] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 862.654520] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882416, 'name': Rename_Task, 'duration_secs': 0.143821} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.655795] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 862.656545] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c25cdc70-fe28-489f-8c4d-e270f99fff36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.659739] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-97bf487f-1771-4294-9223-515654f33784 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.666627] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d38896a-4798-4d43-90d0-404904b9b113 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.670851] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 862.670851] env[62000]: value = "task-882418" [ 862.670851] env[62000]: _type = "Task" [ 862.670851] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.701498] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c29f6b1-fcd3-4493-8ad4-2871d6f868d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.707805] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882418, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.713630] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fdea5d5-cf28-47c2-a050-0229cfe8d50e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.729307] env[62000]: DEBUG nova.compute.provider_tree [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.908211] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.908924] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.909249] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.909552] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.909861] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.910181] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.916176] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0376e5b3-982c-4e19-8b97-69f7e3b26cf4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.934071] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 862.934071] env[62000]: value = "task-882419" [ 862.934071] env[62000]: _type = "Task" [ 862.934071] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.942895] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882419, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.161301] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.161593] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.181416] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882418, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.232795] env[62000]: DEBUG nova.scheduler.client.report [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.443556] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882419, 'name': ReconfigVM_Task, 'duration_secs': 0.18705} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.443884] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 863.664132] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 863.682254] env[62000]: DEBUG oslo_vmware.api [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882418, 'name': PowerOnVM_Task, 'duration_secs': 0.539371} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.683469] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 863.683469] env[62000]: INFO nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Took 7.53 seconds to spawn the instance on the hypervisor. [ 863.683469] env[62000]: DEBUG nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.683868] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc2fa2d-918d-4d67-9a1c-60f633f14d84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.737608] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.358s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.737835] env[62000]: INFO nova.compute.manager [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Migrating [ 863.745355] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.569s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.745619] env[62000]: DEBUG nova.objects.instance [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lazy-loading 'resources' on Instance uuid e9f2cd53-2c73-4c1f-858a-98a291da5547 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 863.951747] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 863.952032] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 863.952032] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 863.952229] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 863.952435] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 863.952604] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 863.952923] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 863.953043] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 863.953263] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 863.953444] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 863.953622] env[62000]: DEBUG nova.virt.hardware [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 863.959402] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfiguring VM instance instance-00000027 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 863.959618] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-596dcf7b-0e16-454b-9248-b270a63b25ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.980882] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 863.980882] env[62000]: value = "task-882420" [ 863.980882] env[62000]: _type = "Task" [ 863.980882] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.989354] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882420, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.190616] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.201935] env[62000]: INFO nova.compute.manager [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Took 30.40 seconds to build instance. [ 864.255455] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.255833] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.256233] env[62000]: DEBUG nova.network.neutron [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.501796] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882420, 'name': ReconfigVM_Task, 'duration_secs': 0.230098} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.502755] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfigured VM instance instance-00000027 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 864.503624] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d121441-ddac-4bc9-b27b-a9ee1c3fb3d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.530567] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfiguring VM instance instance-00000027 to attach disk [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 864.534041] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85d05720-02e3-476c-820d-079ba85e858a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.554058] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 864.554058] env[62000]: value = "task-882421" [ 864.554058] env[62000]: _type = "Task" [ 864.554058] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.568669] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882421, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.588029] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4626f143-11c1-40b6-9081-3e22be23a1ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.596341] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040421bd-5281-4849-aa9f-8f63a6248c94 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.634378] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2699470-2461-40bf-9ea3-ac2b2b2d4433 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.643823] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b54d65fc-54b6-4a58-a13e-f55c29772c20 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.659110] env[62000]: DEBUG nova.compute.provider_tree [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.703998] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e8f8d327-ba6b-4719-8e85-26ace964d78c tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.909s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.998688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.999110] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.999416] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.999652] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.999876] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.002674] env[62000]: INFO nova.compute.manager [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Terminating instance [ 865.004897] env[62000]: DEBUG nova.compute.manager [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 865.005119] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 865.006023] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144ebc89-e8ce-42e0-b130-04a68268ee36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.016763] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 865.017819] env[62000]: DEBUG nova.network.neutron [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.019418] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b3a9f3e5-bc1c-4a8b-8904-3e15b7e23e9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.027538] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 865.027538] env[62000]: value = "task-882422" [ 865.027538] env[62000]: _type = "Task" [ 865.027538] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.036916] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.065577] env[62000]: DEBUG oslo_vmware.api [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882421, 'name': ReconfigVM_Task, 'duration_secs': 0.474065} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.065757] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Reconfigured VM instance instance-00000027 to attach disk [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94/5a8cded8-bcfb-4488-a736-fb6b6aad5a94.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.067032] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 865.162722] env[62000]: DEBUG nova.scheduler.client.report [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.522525] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.537689] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.573345] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-334dd0c9-3c06-4c7c-9b9b-30e0f779a825 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.592333] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda3e3c2-a26a-48a8-beb9-e4c288502ea8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.610126] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 865.645625] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 865.646016] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 865.668336] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.922s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.671058] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.640s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.671360] env[62000]: DEBUG nova.objects.instance [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'resources' on Instance uuid e2e3270d-086b-4441-a3d6-49b05a60b51f {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 865.696632] env[62000]: INFO nova.scheduler.client.report [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Deleted allocations for instance e9f2cd53-2c73-4c1f-858a-98a291da5547 [ 866.041225] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882422, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.152333] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 866.152615] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 866.163643] env[62000]: DEBUG nova.network.neutron [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Port a40f58be-07f1-43ce-afd9-e0ea0c78b634 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 866.207536] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d6c6a1cb-8a7b-473d-bbdb-d72396a2fa80 tempest-ImagesNegativeTestJSON-573056804 tempest-ImagesNegativeTestJSON-573056804-project-member] Lock "e9f2cd53-2c73-4c1f-858a-98a291da5547" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.147s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.440248] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7e1616f-2a64-45b6-98af-a45ecf87e5eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.447985] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-278b51e4-aef9-41ec-a182-ba3e094644fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.478055] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d419e733-a24c-4595-9b82-1b090051456c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.485353] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a170f646-9133-4f9a-bfed-06d0633cb3d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.498846] env[62000]: DEBUG nova.compute.provider_tree [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 866.544803] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882422, 'name': PowerOffVM_Task, 'duration_secs': 1.201897} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.544803] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 866.544803] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 866.545104] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab53e2aa-ad8c-4af5-a360-b59161b432d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.622697] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 866.622933] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 866.623167] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Deleting the datastore file [datastore1] cc7e8af9-16ef-41bf-a91e-edb6c166a3a5 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 866.623459] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b564100-b9af-4f3a-b518-226163b17762 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.629907] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for the task: (returnval){ [ 866.629907] env[62000]: value = "task-882424" [ 866.629907] env[62000]: _type = "Task" [ 866.629907] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.637935] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882424, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.674997] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.675171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquired lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.675341] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Forcefully refreshing network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 867.003438] env[62000]: DEBUG nova.scheduler.client.report [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.050576] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd0c108-157d-410d-9a1f-2781846c7e2f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.069068] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 867.139650] env[62000]: DEBUG oslo_vmware.api [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Task: {'id': task-882424, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.148994} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.139936] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 867.140343] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 867.140492] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 867.140694] env[62000]: INFO nova.compute.manager [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Took 2.14 seconds to destroy the instance on the hypervisor. [ 867.140932] env[62000]: DEBUG oslo.service.loopingcall [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.141144] env[62000]: DEBUG nova.compute.manager [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 867.141247] env[62000]: DEBUG nova.network.neutron [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 867.194282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.194535] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.194725] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.219749] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.508975] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.838s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 867.511625] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.676s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 867.511935] env[62000]: DEBUG nova.objects.instance [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lazy-loading 'resources' on Instance uuid eb27703f-b657-423a-90a9-a7c024a2e473 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 867.522812] env[62000]: DEBUG nova.compute.manager [req-e3e7f20f-ed7e-4835-beaf-dd5c4700e188 req-7f3558b7-b3a7-4f85-933c-ec09ffc2b3a2 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Received event network-vif-deleted-607b5eaf-2995-418d-8242-05869638adf3 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 867.525896] env[62000]: INFO nova.compute.manager [req-e3e7f20f-ed7e-4835-beaf-dd5c4700e188 req-7f3558b7-b3a7-4f85-933c-ec09ffc2b3a2 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Neutron deleted interface 607b5eaf-2995-418d-8242-05869638adf3; detaching it from the instance and deleting it from the info cache [ 867.525896] env[62000]: DEBUG nova.network.neutron [req-e3e7f20f-ed7e-4835-beaf-dd5c4700e188 req-7f3558b7-b3a7-4f85-933c-ec09ffc2b3a2 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.548336] env[62000]: INFO nova.scheduler.client.report [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted allocations for instance e2e3270d-086b-4441-a3d6-49b05a60b51f [ 867.575563] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 867.576174] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0a7c0cf9-df35-42d4-be5f-e381f5dadfe6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.584373] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 867.584373] env[62000]: value = "task-882425" [ 867.584373] env[62000]: _type = "Task" [ 867.584373] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.596458] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882425, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.811494] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.978429] env[62000]: DEBUG nova.network.neutron [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.026654] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0f5bc87-3d95-4580-a1c5-8324c79184db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.045948] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2cdaba-050d-45e5-a3ae-7da02e6255fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.062032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee19f7c2-ff4f-44dd-b38a-f5af0f6f9bfc tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "e2e3270d-086b-4441-a3d6-49b05a60b51f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.474s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.083823] env[62000]: DEBUG nova.compute.manager [req-e3e7f20f-ed7e-4835-beaf-dd5c4700e188 req-7f3558b7-b3a7-4f85-933c-ec09ffc2b3a2 service nova] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Detach interface failed, port_id=607b5eaf-2995-418d-8242-05869638adf3, reason: Instance cc7e8af9-16ef-41bf-a91e-edb6c166a3a5 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 868.099915] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882425, 'name': PowerOffVM_Task, 'duration_secs': 0.268748} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.100260] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 868.100457] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 868.234861] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.235085] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.235310] env[62000]: DEBUG nova.network.neutron [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.268642] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aea5dc1e-1185-4c39-9dda-6ad1855f06ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.276635] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf106cb-2d6e-4357-befe-3037ce27b0cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.307643] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e8ef164-0c81-489e-9edf-4a1e8110d025 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.314688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Releasing lock "refresh_cache-eb27703f-b657-423a-90a9-a7c024a2e473" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.314889] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Updated the network info_cache for instance {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 868.315167] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.316416] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab09be42-48f1-4a98-b630-e851f513dbcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.320303] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.320508] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.321020] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.321244] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.321446] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.321582] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 868.321728] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 868.330312] env[62000]: DEBUG nova.compute.provider_tree [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 868.481543] env[62000]: INFO nova.compute.manager [-] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Took 1.34 seconds to deallocate network for instance. [ 868.606809] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 868.607088] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 868.607257] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 868.607446] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 868.607597] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 868.607749] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 868.607953] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 868.608386] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 868.608601] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 868.608781] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 868.608979] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 868.614058] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-215c2c5b-a34b-4be2-9986-7e9cfd2ba97c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.631540] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 868.631540] env[62000]: value = "task-882426" [ 868.631540] env[62000]: _type = "Task" [ 868.631540] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.639972] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882426, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.833368] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.835057] env[62000]: DEBUG nova.scheduler.client.report [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 868.981748] env[62000]: DEBUG nova.network.neutron [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 868.989961] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 869.141994] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882426, 'name': ReconfigVM_Task, 'duration_secs': 0.287775} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.142358] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 869.339537] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.828s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.342973] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.252s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.344460] env[62000]: INFO nova.compute.claims [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.368082] env[62000]: INFO nova.scheduler.client.report [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Deleted allocations for instance eb27703f-b657-423a-90a9-a7c024a2e473 [ 869.486037] env[62000]: DEBUG oslo_concurrency.lockutils [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.649473] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 869.649742] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 869.650159] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 869.650407] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 869.650596] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 869.650763] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 869.651091] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 869.651240] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 869.653032] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 869.653032] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 869.653032] env[62000]: DEBUG nova.virt.hardware [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 869.657470] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfiguring VM instance instance-00000043 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 869.657909] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b86f1d90-ae4e-4cf6-8b29-3da0b7f7e620 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.680282] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 869.680282] env[62000]: value = "task-882427" [ 869.680282] env[62000]: _type = "Task" [ 869.680282] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.689224] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882427, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.883338] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f59a47d4-9b3e-4b1e-9e62-ab7a840280b3 tempest-VolumesAdminNegativeTest-2122117178 tempest-VolumesAdminNegativeTest-2122117178-project-member] Lock "eb27703f-b657-423a-90a9-a7c024a2e473" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.410s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.008152] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7f5f776-dd4f-4a36-b075-0db7772e65f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.029804] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a909b049-c56a-43e8-824c-62353603d287 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.038178] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 870.182424] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.183431] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9a9d21b-d6e9-48f8-b66e-7c354bd969ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.194943] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882427, 'name': ReconfigVM_Task, 'duration_secs': 0.195362} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.195180] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.195334] env[62000]: ERROR oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk due to incomplete transfer. [ 870.195610] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfigured VM instance instance-00000043 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 870.195848] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-17e1a594-2fe7-4079-9d49-a7c9754588f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.197592] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9979e8-7be9-4480-8a8c-fd4ba2d5edf7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.219991] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.221310] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c1482b98-3d76-4cba-bd58-1760d0e8f56e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.233817] env[62000]: DEBUG oslo_vmware.rw_handles [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52f6a9a3-95c3-e6b3-56f7-078ef2c58dab/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.234031] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Uploaded image 2b9a76bd-b670-46f7-9dda-25750fa1d88d to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 870.236255] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 870.236643] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-357b2af8-210f-43eb-bdb6-e023a9281f4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.242962] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 870.242962] env[62000]: value = "task-882429" [ 870.242962] env[62000]: _type = "Task" [ 870.242962] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.244109] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 870.244109] env[62000]: value = "task-882428" [ 870.244109] env[62000]: _type = "Task" [ 870.244109] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.254254] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882428, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.257182] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882429, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.545016] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-15bb125e-f0af-4fe9-a0b5-fb18d3ca3df8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance '5a8cded8-bcfb-4488-a736-fb6b6aad5a94' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 870.594066] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ed15bb-f7db-4e0d-83c5-14a80eed0da7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.601802] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14936040-8b86-4cf2-8bcb-14354aa7051f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.634793] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c5b952-e699-4bb7-bbf1-1f44e75eb29e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.643366] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab5de2a7-0499-489d-9a55-0a4d14566b1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.655455] env[62000]: DEBUG nova.compute.provider_tree [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 870.660201] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.660422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.756024] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882429, 'name': ReconfigVM_Task, 'duration_secs': 0.323944} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.758899] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b/0a3be3e8-b079-4006-8a46-9b9dd02baa5b.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 870.759221] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 870.763573] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882428, 'name': Destroy_Task, 'duration_secs': 0.362855} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.763804] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Destroyed the VM [ 870.764134] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 870.764293] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4b0c3d23-679e-4fb9-8494-461d400e99aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.777403] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 870.777403] env[62000]: value = "task-882430" [ 870.777403] env[62000]: _type = "Task" [ 870.777403] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.785268] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882430, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.160985] env[62000]: DEBUG nova.scheduler.client.report [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 871.165052] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.268671] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-150da906-6202-462f-96e9-11028d9645fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.294873] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6524acca-f09f-4878-985c-90c753ed5e5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.302718] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882430, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.315587] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 871.672863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.329s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.672863] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 871.677975] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.354s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.679406] env[62000]: INFO nova.compute.claims [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.709827] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.779070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.779070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.779070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.779070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.779070] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.780494] env[62000]: INFO nova.compute.manager [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Terminating instance [ 871.784156] env[62000]: DEBUG nova.compute.manager [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 871.784696] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 871.785703] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17133e26-d282-40b8-90d6-5c5b743e3004 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.799498] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882430, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.801900] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 871.802780] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a68fd678-06c5-43ad-b1e4-1d5a91602dbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.074719] env[62000]: DEBUG nova.network.neutron [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Port 7e199318-132a-4975-92a1-61d761a06165 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 872.180515] env[62000]: DEBUG nova.compute.utils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 872.182138] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 872.182340] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 872.265760] env[62000]: DEBUG nova.policy [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 872.306835] env[62000]: DEBUG oslo_vmware.api [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882430, 'name': RemoveSnapshot_Task, 'duration_secs': 1.476016} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.307115] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 872.307368] env[62000]: INFO nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Took 16.81 seconds to snapshot the instance on the hypervisor. [ 872.317373] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "c354ba41-0915-44cc-9857-42705a56c00f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.317373] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.344608] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 872.344841] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 872.345043] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore1] a432ca35-942d-434a-9cc9-e6e0302a44fd {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.345320] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-df140b9c-c2d4-423d-89d1-9b42dfd670f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.354540] env[62000]: DEBUG oslo_vmware.api [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 872.354540] env[62000]: value = "task-882432" [ 872.354540] env[62000]: _type = "Task" [ 872.354540] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.363111] env[62000]: DEBUG oslo_vmware.api [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882432, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.532908] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Successfully created port: 9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 872.583656] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.583656] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.583656] env[62000]: DEBUG nova.compute.manager [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Going to confirm migration 2 {{(pid=62000) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 872.685732] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 872.814719] env[62000]: DEBUG nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance disappeared during snapshot {{(pid=62000) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 872.822411] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 872.840139] env[62000]: DEBUG nova.compute.manager [None req-29de38c1-c9e4-47cb-97a7-4bf1a7357c25 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image not found during clean up 2b9a76bd-b670-46f7-9dda-25750fa1d88d {{(pid=62000) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 872.867568] env[62000]: DEBUG oslo_vmware.api [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882432, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.212308} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.870848] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 872.871073] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 872.871265] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 872.871456] env[62000]: INFO nova.compute.manager [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Took 1.09 seconds to destroy the instance on the hypervisor. [ 872.871684] env[62000]: DEBUG oslo.service.loopingcall [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 872.872710] env[62000]: DEBUG nova.compute.manager [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 872.873262] env[62000]: DEBUG nova.network.neutron [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.988359] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fadcb17-5b03-476d-b5d2-550eeb176ea4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.998724] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a485299-fa4c-481c-90cb-25d7fabaf0cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.034344] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b66a91-9fc6-4e6e-b263-5de0fffc2ff2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.041864] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6412fabe-4a82-433f-a85e-047e0848f6a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.057191] env[62000]: DEBUG nova.compute.provider_tree [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.102604] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.102872] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.103063] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.185760] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.185958] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.186285] env[62000]: DEBUG nova.network.neutron [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.186537] env[62000]: DEBUG nova.objects.instance [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'info_cache' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.337817] env[62000]: DEBUG nova.compute.manager [req-47908226-67d0-4a62-9c72-78430c8f1125 req-61030b8e-7b68-418f-b833-adeb2d98c672 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Received event network-vif-deleted-01b29391-bf14-4d7b-b13c-8a45618192a2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 873.338091] env[62000]: INFO nova.compute.manager [req-47908226-67d0-4a62-9c72-78430c8f1125 req-61030b8e-7b68-418f-b833-adeb2d98c672 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Neutron deleted interface 01b29391-bf14-4d7b-b13c-8a45618192a2; detaching it from the instance and deleting it from the info cache [ 873.338238] env[62000]: DEBUG nova.network.neutron [req-47908226-67d0-4a62-9c72-78430c8f1125 req-61030b8e-7b68-418f-b833-adeb2d98c672 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.348974] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.560513] env[62000]: DEBUG nova.scheduler.client.report [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.696737] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 873.724177] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 873.724444] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 873.724606] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 873.724791] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 873.724942] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 873.725164] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 873.725390] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 873.725558] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 873.725727] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 873.725893] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 873.726085] env[62000]: DEBUG nova.virt.hardware [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 873.726950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bd1a16-23af-49c2-af97-6a9f46a72530 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.735021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a47bcb4-4e52-49a6-919c-e8d7c741fbc9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.786901] env[62000]: DEBUG nova.network.neutron [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.843020] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eb16d68d-2499-47d7-856c-ff67f552a1ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.852556] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c516c317-ffac-488f-b251-b4867653a692 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.893324] env[62000]: DEBUG nova.compute.manager [req-47908226-67d0-4a62-9c72-78430c8f1125 req-61030b8e-7b68-418f-b833-adeb2d98c672 service nova] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Detach interface failed, port_id=01b29391-bf14-4d7b-b13c-8a45618192a2, reason: Instance a432ca35-942d-434a-9cc9-e6e0302a44fd could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 874.038600] env[62000]: DEBUG nova.compute.manager [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Received event network-vif-plugged-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 874.038600] env[62000]: DEBUG oslo_concurrency.lockutils [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.038600] env[62000]: DEBUG oslo_concurrency.lockutils [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.038600] env[62000]: DEBUG oslo_concurrency.lockutils [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.039191] env[62000]: DEBUG nova.compute.manager [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] No waiting events found dispatching network-vif-plugged-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 874.039535] env[62000]: WARNING nova.compute.manager [req-92c3317b-9beb-49a2-80e9-d526eb905c3f req-55e49109-e61e-475d-a5b5-4ef3c6299320 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Received unexpected event network-vif-plugged-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e for instance with vm_state building and task_state spawning. [ 874.065852] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.066601] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.069251] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.879s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.070856] env[62000]: INFO nova.compute.claims [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.188853] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Successfully updated port: 9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 874.196273] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.196357] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.196752] env[62000]: DEBUG nova.network.neutron [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.290962] env[62000]: INFO nova.compute.manager [-] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Took 1.42 seconds to deallocate network for instance. [ 874.524485] env[62000]: DEBUG nova.network.neutron [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.581695] env[62000]: DEBUG nova.compute.utils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.585113] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.585287] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 874.628548] env[62000]: DEBUG nova.policy [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.695618] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.695697] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.695846] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 874.797610] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.900810] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Successfully created port: f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.941796] env[62000]: DEBUG nova.network.neutron [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.027578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.027814] env[62000]: DEBUG nova.objects.instance [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'migration_context' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 875.085896] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.256873] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.369040] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27933087-0ada-4da8-80c8-ec687da4fdfd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.377408] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2e3152-78bc-4fd3-bce6-6deba1015a7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.414105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbe1f20-1fc8-4e89-93a1-289873607afd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.421893] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d48925-6008-4306-8f2f-1b791d37e601 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.438083] env[62000]: DEBUG nova.compute.provider_tree [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 875.446880] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.453995] env[62000]: DEBUG nova.network.neutron [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updating instance_info_cache with network_info: [{"id": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "address": "fa:16:3e:b0:57:25", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d6e8ee9-bc", "ovs_interfaceid": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.531202] env[62000]: DEBUG nova.objects.base [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Object Instance<5a8cded8-bcfb-4488-a736-fb6b6aad5a94> lazy-loaded attributes: info_cache,migration_context {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 875.532198] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f3ef20-41da-4bab-989e-6f466c8185e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.556893] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d502e64-19b7-4589-ad67-44cf70322a1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.562776] env[62000]: DEBUG oslo_vmware.api [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 875.562776] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5244e34b-2a5d-a506-5ede-a77822b1b6ae" [ 875.562776] env[62000]: _type = "Task" [ 875.562776] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.570443] env[62000]: DEBUG oslo_vmware.api [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5244e34b-2a5d-a506-5ede-a77822b1b6ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.940690] env[62000]: DEBUG nova.scheduler.client.report [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 875.958729] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.958838] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Instance network_info: |[{"id": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "address": "fa:16:3e:b0:57:25", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d6e8ee9-bc", "ovs_interfaceid": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 875.961442] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b0:57:25', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 875.969073] env[62000]: DEBUG oslo.service.loopingcall [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.969784] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 875.970212] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e621d00-b57f-4bec-9813-4b2913efdefc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.986556] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0dcd45-9ad0-4dcf-bd3d-6b42f5ff334c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.008012] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccfbdff-ce94-4371-9f39-c3198d239404 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.010838] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 876.010838] env[62000]: value = "task-882433" [ 876.010838] env[62000]: _type = "Task" [ 876.010838] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.016680] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 876.025070] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882433, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.069805] env[62000]: DEBUG nova.compute.manager [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Received event network-changed-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 876.070053] env[62000]: DEBUG nova.compute.manager [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Refreshing instance network info cache due to event network-changed-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 876.070289] env[62000]: DEBUG oslo_concurrency.lockutils [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] Acquiring lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.070438] env[62000]: DEBUG oslo_concurrency.lockutils [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] Acquired lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.070637] env[62000]: DEBUG nova.network.neutron [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Refreshing network info cache for port 9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 876.081942] env[62000]: DEBUG oslo_vmware.api [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5244e34b-2a5d-a506-5ede-a77822b1b6ae, 'name': SearchDatastore_Task, 'duration_secs': 0.008309} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.082308] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.102237] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.130223] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.130488] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.130661] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.130857] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.131036] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.131181] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.131390] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.131551] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.131785] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.131878] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.132095] env[62000]: DEBUG nova.virt.hardware [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.132984] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696a6218-de2d-4f6f-94f8-80558bd751ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.140711] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f2b23c2-faea-4e9e-a997-ad6a30390f5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.445117] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.376s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.445720] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 876.449035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.616s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.449515] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 876.449716] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 876.450015] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.460s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.450385] env[62000]: DEBUG nova.objects.instance [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lazy-loading 'resources' on Instance uuid cc7e8af9-16ef-41bf-a91e-edb6c166a3a5 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 876.452411] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29b9976-bcce-4b38-bcff-383d5066fb76 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.461788] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ae1ad1-040e-456f-a048-7f44aa11a5fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.484937] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2769b178-5057-4f94-a375-9db6730063b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.492383] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0f3c07-8d45-4c43-955f-e6bd0e4aba3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.524124] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 876.524595] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180530MB free_disk=52GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 876.524731] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.528440] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-828bb932-a115-4b7e-b3b6-fce6df2a2f37 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.536744] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882433, 'name': CreateVM_Task, 'duration_secs': 0.417345} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.537874] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 876.538469] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 876.538469] env[62000]: value = "task-882434" [ 876.538469] env[62000]: _type = "Task" [ 876.538469] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.538815] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.538978] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.539336] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 876.539622] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13e81529-df44-423a-af41-822d42f0bfe7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.547214] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 876.547214] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5264f627-e3cd-8a49-78b2-f2e649262787" [ 876.547214] env[62000]: _type = "Task" [ 876.547214] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.551924] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.565848] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5264f627-e3cd-8a49-78b2-f2e649262787, 'name': SearchDatastore_Task, 'duration_secs': 0.010481} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.565994] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.566227] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 876.566538] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 876.566749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 876.566998] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 876.567367] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd26ea3d-ed04-450d-9004-4d7573c36e4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.579105] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 876.579105] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 876.579105] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-580955a4-2bbb-44fe-9ec3-f73aab442497 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.586019] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 876.586019] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d21147-d848-e2db-7679-b24a0f8af744" [ 876.586019] env[62000]: _type = "Task" [ 876.586019] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.592821] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d21147-d848-e2db-7679-b24a0f8af744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.650563] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Successfully updated port: f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.831874] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.832077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 876.917453] env[62000]: DEBUG nova.network.neutron [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updated VIF entry in instance network info cache for port 9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 876.919611] env[62000]: DEBUG nova.network.neutron [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updating instance_info_cache with network_info: [{"id": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "address": "fa:16:3e:b0:57:25", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9d6e8ee9-bc", "ovs_interfaceid": "9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.953573] env[62000]: DEBUG nova.compute.utils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 876.955140] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 876.955346] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 877.010736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.011087] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.031961] env[62000]: DEBUG nova.policy [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 877.048472] env[62000]: DEBUG oslo_vmware.api [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882434, 'name': PowerOnVM_Task, 'duration_secs': 0.407371} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.050978] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 877.051193] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e338ed73-c77f-4146-8fa0-9e255ab86b23 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance '0a3be3e8-b079-4006-8a46-9b9dd02baa5b' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 877.097453] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d21147-d848-e2db-7679-b24a0f8af744, 'name': SearchDatastore_Task, 'duration_secs': 0.010465} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.101015] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c75694f8-058f-48c5-902f-ce8592c29306 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.107010] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 877.107010] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f4649-8b98-3b4d-79f2-1f2fd4b84988" [ 877.107010] env[62000]: _type = "Task" [ 877.107010] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.122932] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f4649-8b98-3b4d-79f2-1f2fd4b84988, 'name': SearchDatastore_Task, 'duration_secs': 0.00943} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.123293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.123704] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] edc7c99e-d53c-4e0b-991d-c4b693544cdd/edc7c99e-d53c-4e0b-991d-c4b693544cdd.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 877.123822] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f750493d-2f33-493b-a2bc-b09561bfa122 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.133606] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 877.133606] env[62000]: value = "task-882435" [ 877.133606] env[62000]: _type = "Task" [ 877.133606] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.144664] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882435, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.152744] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.152863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.153081] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.246812] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69aba18d-8783-431b-8da2-cb81e22c3212 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.254910] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9886f4c5-459a-4925-97c7-382f503e0670 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.286993] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc620344-0b69-48ad-b21d-ca8252f48aa0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.295335] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f23b7936-8cd5-4747-a488-900334434cb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.311149] env[62000]: DEBUG nova.compute.provider_tree [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 877.338094] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.423231] env[62000]: DEBUG oslo_concurrency.lockutils [req-81eb0fe6-da3a-4672-bb0a-98013c31a256 req-8175dca9-518c-482f-bbab-9e1d5bd12ac1 service nova] Releasing lock "refresh_cache-edc7c99e-d53c-4e0b-991d-c4b693544cdd" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.434649] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Successfully created port: f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 877.458247] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 877.513524] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 877.644124] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882435, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488241} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.644410] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] edc7c99e-d53c-4e0b-991d-c4b693544cdd/edc7c99e-d53c-4e0b-991d-c4b693544cdd.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 877.644643] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.644886] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e5d275f1-e72c-412a-b26c-bb788785f373 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.651175] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 877.651175] env[62000]: value = "task-882436" [ 877.651175] env[62000]: _type = "Task" [ 877.651175] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.660511] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882436, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.708593] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.814406] env[62000]: DEBUG nova.scheduler.client.report [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.868062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.901219] env[62000]: DEBUG nova.network.neutron [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Updating instance_info_cache with network_info: [{"id": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "address": "fa:16:3e:1d:d6:d2", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3b44c3b-72", "ovs_interfaceid": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.010227] env[62000]: INFO nova.compute.manager [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Rebuilding instance [ 878.030648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.052049] env[62000]: DEBUG nova.compute.manager [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 878.052401] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-534ffa1c-9e63-460b-878c-bcf4ca47c67c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.092032] env[62000]: DEBUG nova.compute.manager [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Received event network-vif-plugged-f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.092032] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Acquiring lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.092178] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.092338] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.092508] env[62000]: DEBUG nova.compute.manager [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] No waiting events found dispatching network-vif-plugged-f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 878.092688] env[62000]: WARNING nova.compute.manager [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Received unexpected event network-vif-plugged-f3b44c3b-72c6-43c5-b3a6-30559e284b35 for instance with vm_state building and task_state spawning. [ 878.092843] env[62000]: DEBUG nova.compute.manager [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Received event network-changed-f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 878.093050] env[62000]: DEBUG nova.compute.manager [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Refreshing instance network info cache due to event network-changed-f3b44c3b-72c6-43c5-b3a6-30559e284b35. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 878.093231] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Acquiring lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.164200] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882436, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071209} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.164200] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.164200] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776d8770-e697-4968-92e9-77fe857c3381 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.184921] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] edc7c99e-d53c-4e0b-991d-c4b693544cdd/edc7c99e-d53c-4e0b-991d-c4b693544cdd.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.185182] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d3479d5-3ca1-4ddc-a9f2-ca0e25bc9917 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.205087] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 878.205087] env[62000]: value = "task-882437" [ 878.205087] env[62000]: _type = "Task" [ 878.205087] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.212962] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882437, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.319517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.322658] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.613s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.324953] env[62000]: INFO nova.compute.claims [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 878.345953] env[62000]: INFO nova.scheduler.client.report [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Deleted allocations for instance cc7e8af9-16ef-41bf-a91e-edb6c166a3a5 [ 878.384084] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 878.384377] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.404101] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.404401] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Instance network_info: |[{"id": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "address": "fa:16:3e:1d:d6:d2", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3b44c3b-72", "ovs_interfaceid": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.404680] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Acquired lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.404863] env[62000]: DEBUG nova.network.neutron [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Refreshing network info cache for port f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 878.405985] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:d6:d2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f3b44c3b-72c6-43c5-b3a6-30559e284b35', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.414271] env[62000]: DEBUG oslo.service.loopingcall [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.416816] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 878.417528] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb31bad7-db95-4ec4-93a9-9b47026a8b2c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.438130] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.438130] env[62000]: value = "task-882438" [ 878.438130] env[62000]: _type = "Task" [ 878.438130] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.446658] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882438, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.467294] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.491204] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.491472] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.491634] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.491819] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.492033] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.492222] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.492438] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.492600] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.492772] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.492939] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.493166] env[62000]: DEBUG nova.virt.hardware [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.494007] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a545f49-b1b3-47d6-8302-fecf8c71730c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.502840] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76078f26-d111-4ec0-bc20-f4e7c6b75190 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.572107] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 878.577371] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-201b9089-75d5-4c0e-b36a-5c04530acec1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.586030] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 878.586030] env[62000]: value = "task-882439" [ 878.586030] env[62000]: _type = "Task" [ 878.586030] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.595482] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882439, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.654444] env[62000]: DEBUG nova.network.neutron [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Updated VIF entry in instance network info cache for port f3b44c3b-72c6-43c5-b3a6-30559e284b35. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 878.654889] env[62000]: DEBUG nova.network.neutron [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Updating instance_info_cache with network_info: [{"id": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "address": "fa:16:3e:1d:d6:d2", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf3b44c3b-72", "ovs_interfaceid": "f3b44c3b-72c6-43c5-b3a6-30559e284b35", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.715253] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882437, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.854068] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f0ec2cc9-9f88-46e0-bed8-0900434b8033 tempest-ServerPasswordTestJSON-303676740 tempest-ServerPasswordTestJSON-303676740-project-member] Lock "cc7e8af9-16ef-41bf-a91e-edb6c166a3a5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.855s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.887151] env[62000]: DEBUG nova.compute.utils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.950924] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882438, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.096814] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882439, 'name': PowerOffVM_Task, 'duration_secs': 0.193932} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.097148] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 879.097873] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 879.098160] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8b54b0c2-186c-4895-838c-5ad882b145d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.104906] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 879.104906] env[62000]: value = "task-882440" [ 879.104906] env[62000]: _type = "Task" [ 879.104906] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.115877] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 879.116403] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 879.116403] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201541', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'name': 'volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2f35dab3-6c32-4a35-91ca-1d6859e27f2c', 'attached_at': '', 'detached_at': '', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'serial': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 879.117018] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d5bb76-a8ba-48da-bcc6-df649b898c8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.138367] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80d7b058-2d3c-4bc4-aa1d-397f48f58cf3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.146079] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9216fd23-4316-4122-8f79-1189bd730cd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.168197] env[62000]: DEBUG oslo_concurrency.lockutils [req-63fa0bdd-f5e5-4907-b3a6-749f178db9b0 req-d3794288-c985-4d20-ac57-b134ffadd341 service nova] Releasing lock "refresh_cache-9097dd44-563c-460b-b718-cbc1d5eb4a28" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.169235] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb37982c-b97f-4f6a-97d3-f601d1263064 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.186414] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] The volume has not been displaced from its original location: [datastore2] volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1/volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 879.191694] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Reconfiguring VM instance instance-00000041 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 879.192399] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9532b386-2201-458c-9003-13e4b94c3f5d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.211186] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Successfully updated port: f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 879.213843] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 879.213843] env[62000]: value = "task-882441" [ 879.213843] env[62000]: _type = "Task" [ 879.213843] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.218477] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882437, 'name': ReconfigVM_Task, 'duration_secs': 0.854519} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.221322] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfigured VM instance instance-0000004b to attach disk [datastore2] edc7c99e-d53c-4e0b-991d-c4b693544cdd/edc7c99e-d53c-4e0b-991d-c4b693544cdd.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 879.222376] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9eb73752-415c-4290-8fa5-0a81946ec470 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.230561] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882441, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.231949] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 879.231949] env[62000]: value = "task-882442" [ 879.231949] env[62000]: _type = "Task" [ 879.231949] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.242651] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882442, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.390691] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.450529] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882438, 'name': CreateVM_Task, 'duration_secs': 0.622014} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.453168] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 879.454398] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.454398] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.454653] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 879.454995] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c224cf94-3ea1-43c5-9b9f-73ac8684a218 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.460543] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 879.460543] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c8501-3caf-0ac5-a295-6dcf555043c6" [ 879.460543] env[62000]: _type = "Task" [ 879.460543] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.471256] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c8501-3caf-0ac5-a295-6dcf555043c6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.561730] env[62000]: DEBUG nova.compute.manager [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-vif-plugged-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 879.562117] env[62000]: DEBUG oslo_concurrency.lockutils [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.562246] env[62000]: DEBUG oslo_concurrency.lockutils [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.562426] env[62000]: DEBUG oslo_concurrency.lockutils [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.562569] env[62000]: DEBUG nova.compute.manager [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] No waiting events found dispatching network-vif-plugged-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.562738] env[62000]: WARNING nova.compute.manager [req-4ac74104-a208-413c-995c-8a88f482fecc req-ad5ba89b-eb44-4460-b4ad-0782683d8f2d service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received unexpected event network-vif-plugged-f0d1137e-2373-4645-85db-8c5c8c963ff4 for instance with vm_state building and task_state spawning. [ 879.630925] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb938a9-32cf-40e7-8217-c6a5f11d7e41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.638488] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33eab99f-3e8a-4bce-97fb-d21a9a06e563 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.669524] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.669901] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.670145] env[62000]: DEBUG nova.compute.manager [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Going to confirm migration 3 {{(pid=62000) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 879.672331] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac08a16-d969-4999-a13c-ea7b12334f30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.681590] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-190baa25-3523-4f55-acbe-6d3c9b3a0a59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.697867] env[62000]: DEBUG nova.compute.provider_tree [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 879.721017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.721017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.721017] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.734018] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882441, 'name': ReconfigVM_Task, 'duration_secs': 0.171057} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.734018] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Reconfigured VM instance instance-00000041 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 879.741498] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c89a0b5b-85cb-4ab4-b810-84c6bd173c7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.757930] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882442, 'name': Rename_Task, 'duration_secs': 0.245759} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.758325] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 879.759600] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cdc4d1b8-ddbc-48f0-9714-d3dbdf0c8311 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.761449] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 879.761449] env[62000]: value = "task-882443" [ 879.761449] env[62000]: _type = "Task" [ 879.761449] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.770971] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 879.770971] env[62000]: value = "task-882444" [ 879.770971] env[62000]: _type = "Task" [ 879.770971] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.774613] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882443, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.780160] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882444, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.976763] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520c8501-3caf-0ac5-a295-6dcf555043c6, 'name': SearchDatastore_Task, 'duration_secs': 0.010876} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.977290] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.977754] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.978433] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.978849] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.979358] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.980121] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4ac02d3-fa91-4a09-ac51-2d1c8e3a3582 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.992021] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.992021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 879.992021] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64029a82-8289-4b3c-a130-a45820545e2e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.997222] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 879.997222] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959fae-e180-ea2b-7a53-54f0488503ad" [ 879.997222] env[62000]: _type = "Task" [ 879.997222] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.006313] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959fae-e180-ea2b-7a53-54f0488503ad, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.201200] env[62000]: DEBUG nova.scheduler.client.report [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 880.221659] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.221863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.222139] env[62000]: DEBUG nova.network.neutron [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 880.222405] env[62000]: DEBUG nova.objects.instance [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'info_cache' on Instance uuid 0a3be3e8-b079-4006-8a46-9b9dd02baa5b {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 880.270935] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882443, 'name': ReconfigVM_Task, 'duration_secs': 0.199406} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.274782] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.276825] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201541', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'name': 'volume-2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '2f35dab3-6c32-4a35-91ca-1d6859e27f2c', 'attached_at': '', 'detached_at': '', 'volume_id': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1', 'serial': '2b463ae3-a1c9-43a7-a945-3da06e9ebae1'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 880.276955] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 880.278013] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad6ad929-da23-4a98-8ee2-c91f86e7eda3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.285947] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882444, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.287998] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 880.288609] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0bf3ea4a-931a-4d9d-8058-ef01ca6f2876 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.374881] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 880.375280] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 880.375336] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Deleting the datastore file [datastore2] 2f35dab3-6c32-4a35-91ca-1d6859e27f2c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 880.375956] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-28adf9dc-4087-482a-a125-d08526affe92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.383504] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for the task: (returnval){ [ 880.383504] env[62000]: value = "task-882446" [ 880.383504] env[62000]: _type = "Task" [ 880.383504] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.392503] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882446, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.470585] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.470958] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.471302] env[62000]: INFO nova.compute.manager [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Attaching volume acf81f65-f409-4b80-9f5c-10c061b9ad88 to /dev/sdb [ 880.508621] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52959fae-e180-ea2b-7a53-54f0488503ad, 'name': SearchDatastore_Task, 'duration_secs': 0.011214} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.509423] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0386d4f-e0e0-4198-97bf-62983501e154 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.514955] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 880.514955] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52351df0-8c1d-f6e8-4228-0bb7108b9326" [ 880.514955] env[62000]: _type = "Task" [ 880.514955] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.516618] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4348ad2f-84a8-4282-bfc3-7a0cfcf8f4c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.531080] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52351df0-8c1d-f6e8-4228-0bb7108b9326, 'name': SearchDatastore_Task, 'duration_secs': 0.008951} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.531473] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.533113] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9097dd44-563c-460b-b718-cbc1d5eb4a28/9097dd44-563c-460b-b718-cbc1d5eb4a28.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 880.533113] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2025e33-4a92-41cf-bf5a-157e571dc541 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.535493] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cf09877d-ef7c-470a-b50b-37660ae31bb0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.541835] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 880.541835] env[62000]: value = "task-882447" [ 880.541835] env[62000]: _type = "Task" [ 880.541835] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.553948] env[62000]: DEBUG nova.virt.block_device [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating existing volume attachment record: 53e70b02-d94e-4690-aa9b-ad3b56d7c7d9 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 880.566161] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882447, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.599922] env[62000]: DEBUG nova.network.neutron [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.707754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.708557] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 880.712245] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.362s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.713321] env[62000]: INFO nova.compute.claims [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.782891] env[62000]: DEBUG oslo_vmware.api [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882444, 'name': PowerOnVM_Task, 'duration_secs': 0.758913} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.783235] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 880.783560] env[62000]: INFO nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Took 7.09 seconds to spawn the instance on the hypervisor. [ 880.783950] env[62000]: DEBUG nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 880.785066] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a05720-0227-4f17-b285-d8e700a954b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.895743] env[62000]: DEBUG oslo_vmware.api [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Task: {'id': task-882446, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.096841} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.896158] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 880.896409] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 880.896780] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 880.969855] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 880.970342] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ff2ca07-21f3-4d3e-9d10-37461d26237d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.980741] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89798d77-b14d-4485-b358-53278026d32d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.015422] env[62000]: ERROR nova.compute.manager [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Failed to detach volume 2b463ae3-a1c9-43a7-a945-3da06e9ebae1 from /dev/sda: nova.exception.InstanceNotFound: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Traceback (most recent call last): [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self.driver.rebuild(**kwargs) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise NotImplementedError() [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] NotImplementedError [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] During handling of the above exception, another exception occurred: [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Traceback (most recent call last): [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self.driver.detach_volume(context, old_connection_info, [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] return self._volumeops.detach_volume(connection_info, instance) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._detach_volume_vmdk(connection_info, instance) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] stable_ref.fetch_moref(session) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise exception.InstanceNotFound(instance_id=self._uuid) [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] nova.exception.InstanceNotFound: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. [ 881.015422] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.053522] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882447, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.105023] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.106131] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Instance network_info: |[{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.106131] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7d:b4:87', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a15de394-0367-4921-a5c1-6ac8615e3283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f0d1137e-2373-4645-85db-8c5c8c963ff4', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.115644] env[62000]: DEBUG oslo.service.loopingcall [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.119605] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 881.120707] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6674b3b7-f153-4e2e-b184-50c025ad41d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.147777] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.147777] env[62000]: value = "task-882451" [ 881.147777] env[62000]: _type = "Task" [ 881.147777] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.156325] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882451, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.198121] env[62000]: DEBUG nova.compute.utils [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Build of instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c aborted: Failed to rebuild volume backed instance. {{(pid=62000) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 881.200811] env[62000]: ERROR nova.compute.manager [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c aborted: Failed to rebuild volume backed instance. [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Traceback (most recent call last): [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self.driver.rebuild(**kwargs) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise NotImplementedError() [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] NotImplementedError [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] During handling of the above exception, another exception occurred: [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Traceback (most recent call last): [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._detach_root_volume(context, instance, root_bdm) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] with excutils.save_and_reraise_exception(): [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self.force_reraise() [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise self.value [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self.driver.detach_volume(context, old_connection_info, [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 553, in detach_volume [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] return self._volumeops.detach_volume(connection_info, instance) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._detach_volume_vmdk(connection_info, instance) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1135, in get_vm_ref [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] stable_ref.fetch_moref(session) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1126, in fetch_moref [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise exception.InstanceNotFound(instance_id=self._uuid) [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] nova.exception.InstanceNotFound: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] During handling of the above exception, another exception occurred: [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Traceback (most recent call last): [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] yield [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 881.200811] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._do_rebuild_instance_with_claim( [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._do_rebuild_instance( [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._rebuild_default_impl(**kwargs) [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] self._rebuild_volume_backed_instance( [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] raise exception.BuildAbortException( [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] nova.exception.BuildAbortException: Build of instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c aborted: Failed to rebuild volume backed instance. [ 881.202917] env[62000]: ERROR nova.compute.manager [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] [ 881.218815] env[62000]: DEBUG nova.compute.utils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 881.223297] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 881.223439] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 881.305517] env[62000]: DEBUG nova.policy [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79730e91610c4c598019b122b4f41418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6106a70abdad4111977eebbd82434337', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 881.309368] env[62000]: INFO nova.compute.manager [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Took 26.24 seconds to build instance. [ 881.550258] env[62000]: DEBUG nova.network.neutron [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [{"id": "7e199318-132a-4975-92a1-61d761a06165", "address": "fa:16:3e:a2:d6:44", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e199318-13", "ovs_interfaceid": "7e199318-132a-4975-92a1-61d761a06165", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.562861] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882447, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.597647} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.564284] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9097dd44-563c-460b-b718-cbc1d5eb4a28/9097dd44-563c-460b-b718-cbc1d5eb4a28.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 881.564695] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.565537] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c471d071-11e3-443d-8859-bfcb64e33a7f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.574206] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 881.574206] env[62000]: value = "task-882452" [ 881.574206] env[62000]: _type = "Task" [ 881.574206] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.582920] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882452, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.656950] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882451, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.712941] env[62000]: DEBUG nova.compute.manager [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 881.713367] env[62000]: DEBUG nova.compute.manager [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 881.713457] env[62000]: DEBUG oslo_concurrency.lockutils [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.713606] env[62000]: DEBUG oslo_concurrency.lockutils [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 881.713771] env[62000]: DEBUG nova.network.neutron [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 881.724415] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.810981] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d95a5695-f14e-4483-9da8-a8449423d2dc tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.750s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.817797] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Successfully created port: 919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 882.021888] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c7a0e40-3e3c-4a8f-9e13-4e4d78236d47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.032089] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53352889-1302-4ea5-9aff-14749a331eaa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.063679] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-0a3be3e8-b079-4006-8a46-9b9dd02baa5b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.063951] env[62000]: DEBUG nova.objects.instance [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'migration_context' on Instance uuid 0a3be3e8-b079-4006-8a46-9b9dd02baa5b {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.065145] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.065372] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.065545] env[62000]: INFO nova.compute.manager [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Shelving [ 882.067735] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02d6cc16-2547-4a40-9523-e213e3827e1a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.079495] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b530dd29-6bd1-44a8-a641-10392113e1cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.089641] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882452, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062453} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.098229] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.099653] env[62000]: DEBUG nova.compute.provider_tree [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.100751] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-216114a0-e1a1-4f6a-875f-8aa2b8f6dbd0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.130348] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 9097dd44-563c-460b-b718-cbc1d5eb4a28/9097dd44-563c-460b-b718-cbc1d5eb4a28.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.130348] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61419e37-d791-4040-a83f-0a78e15e438c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.153044] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 882.153044] env[62000]: value = "task-882453" [ 882.153044] env[62000]: _type = "Task" [ 882.153044] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.159195] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882451, 'name': CreateVM_Task, 'duration_secs': 0.526943} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.159687] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 882.160394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.160595] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.160921] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.162106] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a94ce65b-02c1-4d74-a5dd-e86d94a720f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.165909] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882453, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.168944] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 882.168944] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529b47e1-c530-a78f-a7b9-31e27385532c" [ 882.168944] env[62000]: _type = "Task" [ 882.168944] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.176746] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529b47e1-c530-a78f-a7b9-31e27385532c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.415029] env[62000]: DEBUG nova.network.neutron [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 882.415465] env[62000]: DEBUG nova.network.neutron [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.570979] env[62000]: DEBUG nova.objects.base [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Object Instance<0a3be3e8-b079-4006-8a46-9b9dd02baa5b> lazy-loaded attributes: info_cache,migration_context {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 882.573783] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca7df97-d454-4006-a1d9-1b41a4b9ce37 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.576640] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 882.576877] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4702b053-db13-4606-a36a-8f08d80fd77c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.598163] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-221390db-3950-4673-b656-a62846d24085 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.600751] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 882.600751] env[62000]: value = "task-882454" [ 882.600751] env[62000]: _type = "Task" [ 882.600751] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.606456] env[62000]: DEBUG nova.scheduler.client.report [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 882.609779] env[62000]: DEBUG oslo_vmware.api [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 882.609779] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b713c2-060a-5ab4-373f-6320d9a832ed" [ 882.609779] env[62000]: _type = "Task" [ 882.609779] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.616487] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882454, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.621469] env[62000]: DEBUG oslo_vmware.api [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b713c2-060a-5ab4-373f-6320d9a832ed, 'name': SearchDatastore_Task, 'duration_secs': 0.006157} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.622409] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.663535] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882453, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.678555] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529b47e1-c530-a78f-a7b9-31e27385532c, 'name': SearchDatastore_Task, 'duration_secs': 0.010763} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.678876] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.679133] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 882.679376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.679526] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.679710] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 882.679975] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7d67941f-1512-4096-bc41-bdf18b283671 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.688072] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 882.688267] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 882.688943] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-856bea1b-07c1-4d22-a167-65fff39afde9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.694343] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 882.694343] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a19156-d8fe-1ab1-2710-f53d0d959522" [ 882.694343] env[62000]: _type = "Task" [ 882.694343] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.702480] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a19156-d8fe-1ab1-2710-f53d0d959522, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.739640] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 882.768472] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.768753] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.770030] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.770314] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.770388] env[62000]: DEBUG nova.virt.hardware [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.771661] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e43511de-a4c8-4352-a289-c922b98375c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.779264] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bb2a47-2058-46aa-93fc-f0cce0ea160e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.918718] env[62000]: DEBUG oslo_concurrency.lockutils [req-09e56a5e-391e-4765-8e15-b9e192bcaae3 req-c97d2c87-28e5-4b39-bf6a-6687e0c79746 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.030581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.030843] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.112374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.112952] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 883.115711] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882454, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.116013] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.319s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.116263] env[62000]: DEBUG nova.objects.instance [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'resources' on Instance uuid a432ca35-942d-434a-9cc9-e6e0302a44fd {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.164430] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882453, 'name': ReconfigVM_Task, 'duration_secs': 0.920164} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.164754] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 9097dd44-563c-460b-b718-cbc1d5eb4a28/9097dd44-563c-460b-b718-cbc1d5eb4a28.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.165393] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6cb2f963-dbb7-436a-9586-95ca1a28176a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.172177] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 883.172177] env[62000]: value = "task-882456" [ 883.172177] env[62000]: _type = "Task" [ 883.172177] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.181872] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882456, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.204678] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a19156-d8fe-1ab1-2710-f53d0d959522, 'name': SearchDatastore_Task, 'duration_secs': 0.010074} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.205484] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22755264-089b-4e69-96ba-261509c986da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.210732] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 883.210732] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528d113e-797d-8372-841a-91c9ec06d6f0" [ 883.210732] env[62000]: _type = "Task" [ 883.210732] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.219635] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.219932] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528d113e-797d-8372-841a-91c9ec06d6f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.534563] env[62000]: DEBUG nova.compute.utils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.611476] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882454, 'name': PowerOffVM_Task, 'duration_secs': 0.563241} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.611763] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 883.612648] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17088d38-7c49-49bd-9088-af8aee176634 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.630787] env[62000]: DEBUG nova.compute.utils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 883.635807] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 883.635990] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 883.638443] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e223b939-6095-4806-9e02-0383e2c6614e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.683384] env[62000]: DEBUG nova.policy [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffd8327dd8c4844b776982e99de5e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14aec090f2a4abc8078df1fe4428bbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 883.685129] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882456, 'name': Rename_Task, 'duration_secs': 0.166634} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.687485] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 883.688341] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ec068164-2e8b-4e41-8c7f-282eda570fe7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.695409] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 883.695409] env[62000]: value = "task-882457" [ 883.695409] env[62000]: _type = "Task" [ 883.695409] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.703511] env[62000]: DEBUG nova.compute.manager [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Received event network-vif-plugged-919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 883.703752] env[62000]: DEBUG oslo_concurrency.lockutils [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.703995] env[62000]: DEBUG oslo_concurrency.lockutils [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.704207] env[62000]: DEBUG oslo_concurrency.lockutils [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.704387] env[62000]: DEBUG nova.compute.manager [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] No waiting events found dispatching network-vif-plugged-919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.704559] env[62000]: WARNING nova.compute.manager [req-a60654d2-3fbc-48dd-b0f1-31a68647bf20 req-73b2ae1f-776f-4348-981c-466682bef9d6 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Received unexpected event network-vif-plugged-919b3f5d-af5d-45e1-896f-ac0ea1309017 for instance with vm_state building and task_state spawning. [ 883.711179] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882457, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.721401] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528d113e-797d-8372-841a-91c9ec06d6f0, 'name': SearchDatastore_Task, 'duration_secs': 0.010381} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.722030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.722346] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 72a5bf70-dc6e-4887-abb8-8fbad64bb065/72a5bf70-dc6e-4887-abb8-8fbad64bb065.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 883.722652] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-675eac11-9978-4ce2-a612-023e344157bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.724811] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.725090] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.725327] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.725558] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.725755] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.732826] env[62000]: INFO nova.compute.manager [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Terminating instance [ 883.735023] env[62000]: DEBUG nova.compute.manager [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.735927] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d60dee16-2fc1-4bb3-b7a0-c53c955a9629 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.738262] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 883.738262] env[62000]: value = "task-882458" [ 883.738262] env[62000]: _type = "Task" [ 883.738262] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.745576] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3924bd9-5557-4d91-899a-9247b9bf6aa5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.761991] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882458, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.783914] env[62000]: WARNING nova.virt.vmwareapi.driver [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. [ 883.784168] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 883.784509] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c21f07bb-1dd3-46b9-8826-0922db1ebad2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.792863] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-532ef87b-54f7-42f6-b59c-df24b7d183fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.828366] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. [ 883.828619] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 883.828823] env[62000]: INFO nova.compute.manager [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Took 0.09 seconds to destroy the instance on the hypervisor. [ 883.829113] env[62000]: DEBUG oslo.service.loopingcall [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.829360] env[62000]: DEBUG nova.compute.manager [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.829465] env[62000]: DEBUG nova.network.neutron [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 883.939573] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-402482ea-cdee-46c8-ba0a-71676eca8ed6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.948396] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab11498b-51dc-4978-8419-e92cbf2abdb8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.985373] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Successfully updated port: 919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 883.990155] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af777aa-c9e9-4c91-965c-b10fee9a27f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.998267] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f14cc2-56ed-430b-8c0b-db5713977569 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.015823] env[62000]: DEBUG nova.compute.provider_tree [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.037424] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.136759] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 884.151032] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 884.151032] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-580db8e2-e12f-4350-87d8-99d1177fce47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.160980] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 884.160980] env[62000]: value = "task-882459" [ 884.160980] env[62000]: _type = "Task" [ 884.160980] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.173303] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882459, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.209563] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882457, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.249529] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882458, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.300044] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Successfully created port: e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 884.366523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.366804] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.367037] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.367259] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.367459] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.369728] env[62000]: INFO nova.compute.manager [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Terminating instance [ 884.371541] env[62000]: DEBUG nova.compute.manager [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 884.371741] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 884.372759] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57da4ba7-9b89-4859-bfee-6233d07d33aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.381608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 884.381859] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9b00c767-3497-4467-9669-af580089d68e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.389438] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 884.389438] env[62000]: value = "task-882460" [ 884.389438] env[62000]: _type = "Task" [ 884.389438] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.397102] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882460, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.478663] env[62000]: DEBUG nova.compute.manager [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Received event network-changed-919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.478840] env[62000]: DEBUG nova.compute.manager [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Refreshing instance network info cache due to event network-changed-919b3f5d-af5d-45e1-896f-ac0ea1309017. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 884.479082] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] Acquiring lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.479205] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] Acquired lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 884.479375] env[62000]: DEBUG nova.network.neutron [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Refreshing network info cache for port 919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 884.492065] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.519263] env[62000]: DEBUG nova.scheduler.client.report [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.671249] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882459, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.706243] env[62000]: DEBUG oslo_vmware.api [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882457, 'name': PowerOnVM_Task, 'duration_secs': 0.863201} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.706515] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 884.706720] env[62000]: INFO nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Took 8.60 seconds to spawn the instance on the hypervisor. [ 884.706906] env[62000]: DEBUG nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 884.707683] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b499f12-b4f7-4737-ac37-50dbd95406fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.749628] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882458, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.591846} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.749928] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 72a5bf70-dc6e-4887-abb8-8fbad64bb065/72a5bf70-dc6e-4887-abb8-8fbad64bb065.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 884.750124] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.750378] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b11e3a75-a992-4e65-80de-473f042b0d3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.757473] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 884.757473] env[62000]: value = "task-882461" [ 884.757473] env[62000]: _type = "Task" [ 884.757473] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.766215] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882461, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.901842] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882460, 'name': PowerOffVM_Task, 'duration_secs': 0.171873} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.902188] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 884.902493] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 884.902763] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ec131ab0-985d-4810-9186-6c59e28005db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.907203] env[62000]: DEBUG nova.compute.manager [req-9a9c9dfb-9548-47ab-add3-eac6553cc828 req-0700f914-99ab-46b9-b948-68e97e8ec6a3 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Received event network-vif-deleted-3c405e0f-e997-4845-a292-10cad3943d50 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 884.907400] env[62000]: INFO nova.compute.manager [req-9a9c9dfb-9548-47ab-add3-eac6553cc828 req-0700f914-99ab-46b9-b948-68e97e8ec6a3 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Neutron deleted interface 3c405e0f-e997-4845-a292-10cad3943d50; detaching it from the instance and deleting it from the info cache [ 884.907583] env[62000]: DEBUG nova.network.neutron [req-9a9c9dfb-9548-47ab-add3-eac6553cc828 req-0700f914-99ab-46b9-b948-68e97e8ec6a3 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.010319] env[62000]: DEBUG nova.network.neutron [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.026111] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.030950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 8.948s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.061148] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 885.061148] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 885.061148] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Deleting the datastore file [datastore2] 4e3f804c-f4a2-44ab-bc84-3c97d5803fab {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 885.061356] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9be9eaaa-ec76-40a6-971a-31fb3d8186ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.067295] env[62000]: INFO nova.scheduler.client.report [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted allocations for instance a432ca35-942d-434a-9cc9-e6e0302a44fd [ 885.078607] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for the task: (returnval){ [ 885.078607] env[62000]: value = "task-882463" [ 885.078607] env[62000]: _type = "Task" [ 885.078607] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.088203] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882463, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.122265] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.122265] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.122265] env[62000]: INFO nova.compute.manager [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Attaching volume 3e32054b-1f06-4121-8f0c-31b7d7ea244a to /dev/sdb [ 885.125055] env[62000]: DEBUG nova.network.neutron [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.139410] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 885.139546] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201587', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'name': 'volume-acf81f65-f409-4b80-9f5c-10c061b9ad88', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66edb906-f936-4f57-833d-224f36af109e', 'attached_at': '', 'detached_at': '', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'serial': 'acf81f65-f409-4b80-9f5c-10c061b9ad88'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 885.142127] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1336e87-7623-4fc1-b61d-cf77d6435454 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.148988] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 885.172539] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f125af-e9a6-4ecb-8e6a-cba6a3bc41f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.175908] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bddf1de-f3fb-4f8b-860f-41dd47e81ad0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.184955] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882459, 'name': CreateSnapshot_Task, 'duration_secs': 0.719398} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.199650] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 885.207829] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] volume-acf81f65-f409-4b80-9f5c-10c061b9ad88/volume-acf81f65-f409-4b80-9f5c-10c061b9ad88.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.210761] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f864b3-b137-4aea-b51e-ffda7b2bc5dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.214087] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997ea842-34c7-46da-8fd8-24c714850599 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.217451] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b262210f-c8e9-43f7-a6bc-41b71f0e01f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.237932] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 885.237932] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 885.238105] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 885.238308] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 885.238474] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 885.238698] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 885.238967] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 885.239202] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 885.239459] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 885.239618] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 885.239848] env[62000]: DEBUG nova.virt.hardware [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 885.241593] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80688299-36a2-4a23-974b-53266073d3ed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.246463] env[62000]: INFO nova.compute.manager [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Took 22.94 seconds to build instance. [ 885.251297] env[62000]: DEBUG nova.virt.block_device [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updating existing volume attachment record: cf0f565f-e536-4147-afd1-3624522c7918 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 885.259405] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 885.259405] env[62000]: value = "task-882464" [ 885.259405] env[62000]: _type = "Task" [ 885.259405] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.268020] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee54f399-b973-478d-a749-c719275fecfb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.295387] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882464, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.295387] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882461, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.077366} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.295387] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.295387] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-163cc342-6e53-4c0d-9426-7adedd1f1bf2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.321717] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] 72a5bf70-dc6e-4887-abb8-8fbad64bb065/72a5bf70-dc6e-4887-abb8-8fbad64bb065.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.321717] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-115589ae-770d-495a-ad7f-17be33c5555c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.342667] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 885.342667] env[62000]: value = "task-882465" [ 885.342667] env[62000]: _type = "Task" [ 885.342667] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.351032] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882465, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.375766] env[62000]: DEBUG nova.network.neutron [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.411232] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4ce3852-d200-440a-b08c-d877a7a7859a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.426797] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2afb1820-c88c-4f23-92e4-f6a2228f4616 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.457382] env[62000]: DEBUG nova.compute.manager [req-9a9c9dfb-9548-47ab-add3-eac6553cc828 req-0700f914-99ab-46b9-b948-68e97e8ec6a3 service nova] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Detach interface failed, port_id=3c405e0f-e997-4845-a292-10cad3943d50, reason: Instance 2f35dab3-6c32-4a35-91ca-1d6859e27f2c could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 885.584496] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c292082e-3a3a-444e-ba4b-d324f027167c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "a432ca35-942d-434a-9cc9-e6e0302a44fd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.807s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.597698] env[62000]: DEBUG oslo_vmware.api [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Task: {'id': task-882463, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161568} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.597698] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 885.597698] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 885.597698] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 885.597698] env[62000]: INFO nova.compute.manager [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Took 1.22 seconds to destroy the instance on the hypervisor. [ 885.597698] env[62000]: DEBUG oslo.service.loopingcall [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 885.597698] env[62000]: DEBUG nova.compute.manager [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 885.597698] env[62000]: DEBUG nova.network.neutron [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 885.633985] env[62000]: DEBUG oslo_concurrency.lockutils [req-6a4a9931-0c1b-449f-b650-99f89b081aaa req-67e94166-48a9-4988-aebb-31e687fbfe0b service nova] Releasing lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.635331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.635331] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 885.748484] env[62000]: DEBUG oslo_concurrency.lockutils [None req-14d8a487-a55b-4099-996a-a6b92cbd75bb tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.454s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.787068] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 885.787787] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882464, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.790713] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-bd074464-912a-4a13-ba99-3c159d87809e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.800735] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 885.800735] env[62000]: value = "task-882469" [ 885.800735] env[62000]: _type = "Task" [ 885.800735] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.811426] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882469, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.858393] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882465, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.860250] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39a5a794-fdbb-4de7-83c9-bd32766d0ab2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.868393] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10f86a84-3bc5-4a7b-aecb-d2afa4fb9746 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.899022] env[62000]: INFO nova.compute.manager [-] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Took 2.07 seconds to deallocate network for instance. [ 885.902330] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd77c7d-6328-409f-b162-a10f8b066a05 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.913932] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be95ac6-cd39-40e0-a909-2c84e581c814 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.920105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.920950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.925064] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.925064] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.925064] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.926707] env[62000]: INFO nova.compute.manager [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Terminating instance [ 885.930475] env[62000]: DEBUG nova.compute.manager [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 885.930831] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 885.933480] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f2fe8c-fb71-462b-bdac-0e87f1da4618 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.954226] env[62000]: DEBUG nova.compute.provider_tree [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 885.964305] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 885.964696] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7ce71a3e-8b7a-4c96-9f12-b5827d4321dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.974868] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 885.974868] env[62000]: value = "task-882470" [ 885.974868] env[62000]: _type = "Task" [ 885.974868] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.988886] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882470, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.155673] env[62000]: DEBUG nova.compute.manager [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Received event network-vif-plugged-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.155673] env[62000]: DEBUG oslo_concurrency.lockutils [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] Acquiring lock "c354ba41-0915-44cc-9857-42705a56c00f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.155869] env[62000]: DEBUG oslo_concurrency.lockutils [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] Lock "c354ba41-0915-44cc-9857-42705a56c00f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.156346] env[62000]: DEBUG oslo_concurrency.lockutils [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] Lock "c354ba41-0915-44cc-9857-42705a56c00f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.156431] env[62000]: DEBUG nova.compute.manager [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] No waiting events found dispatching network-vif-plugged-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.156663] env[62000]: WARNING nova.compute.manager [req-45994bbd-c9b3-4ec4-87f3-d70ab5a61578 req-93a30a8f-d9d6-40cc-9705-735c237cf244 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Received unexpected event network-vif-plugged-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 for instance with vm_state building and task_state spawning. [ 886.193306] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.272618] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882464, 'name': ReconfigVM_Task, 'duration_secs': 0.679996} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.272930] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfigured VM instance instance-00000040 to attach disk [datastore1] volume-acf81f65-f409-4b80-9f5c-10c061b9ad88/volume-acf81f65-f409-4b80-9f5c-10c061b9ad88.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.279975] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5486528c-5383-424f-b9f3-fcedfafcbf29 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.297082] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 886.297082] env[62000]: value = "task-882471" [ 886.297082] env[62000]: _type = "Task" [ 886.297082] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.309756] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882469, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.312861] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882471, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.354583] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882465, 'name': ReconfigVM_Task, 'duration_secs': 0.854423} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.354957] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfigured VM instance instance-0000004d to attach disk [datastore1] 72a5bf70-dc6e-4887-abb8-8fbad64bb065/72a5bf70-dc6e-4887-abb8-8fbad64bb065.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.356500] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ce72ef3-24ba-4966-8c05-f2febc04b0d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.365703] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 886.365703] env[62000]: value = "task-882472" [ 886.365703] env[62000]: _type = "Task" [ 886.365703] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.375979] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882472, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.439070] env[62000]: DEBUG nova.network.neutron [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating instance_info_cache with network_info: [{"id": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "address": "fa:16:3e:86:36:44", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919b3f5d-af", "ovs_interfaceid": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.462681] env[62000]: DEBUG nova.scheduler.client.report [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.474587] env[62000]: INFO nova.compute.manager [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Took 0.57 seconds to detach 1 volumes for instance. [ 886.479924] env[62000]: DEBUG nova.compute.manager [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Deleting volume: 2b463ae3-a1c9-43a7-a945-3da06e9ebae1 {{(pid=62000) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 886.492847] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882470, 'name': PowerOffVM_Task, 'duration_secs': 0.266548} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.493328] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 886.493611] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 886.493896] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50214e1c-2f97-420a-bc06-1f5e54a0ec39 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.735938] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 886.736247] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 886.736393] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore2] 9097dd44-563c-460b-b718-cbc1d5eb4a28 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 886.736924] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e31709aa-0d02-4d41-b29a-62ac9666a46c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.748030] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 886.748030] env[62000]: value = "task-882475" [ 886.748030] env[62000]: _type = "Task" [ 886.748030] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.751372] env[62000]: DEBUG nova.network.neutron [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.755621] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882475, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.782727] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Successfully updated port: e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 886.814292] env[62000]: DEBUG oslo_vmware.api [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882471, 'name': ReconfigVM_Task, 'duration_secs': 0.144643} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.816938] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201587', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'name': 'volume-acf81f65-f409-4b80-9f5c-10c061b9ad88', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66edb906-f936-4f57-833d-224f36af109e', 'attached_at': '', 'detached_at': '', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'serial': 'acf81f65-f409-4b80-9f5c-10c061b9ad88'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 886.817911] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882469, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.876250] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882472, 'name': Rename_Task, 'duration_secs': 0.26653} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.876634] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 886.877142] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b77b602d-6156-48f7-ae75-d60f32e129a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.883333] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 886.883333] env[62000]: value = "task-882476" [ 886.883333] env[62000]: _type = "Task" [ 886.883333] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.893657] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882476, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.941865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.942386] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance network_info: |[{"id": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "address": "fa:16:3e:86:36:44", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919b3f5d-af", "ovs_interfaceid": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 886.942871] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:36:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '919b3f5d-af5d-45e1-896f-ac0ea1309017', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 886.957787] env[62000]: DEBUG oslo.service.loopingcall [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 886.958068] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 886.958324] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-084fdc68-8d33-4ece-8980-6352d439ff99 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.983893] env[62000]: DEBUG nova.compute.manager [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Received event network-vif-deleted-7801caca-efe4-49ff-b583-7e91be8cf888 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.984236] env[62000]: DEBUG nova.compute.manager [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Received event network-changed-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 886.984446] env[62000]: DEBUG nova.compute.manager [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Refreshing instance network info cache due to event network-changed-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 886.984514] env[62000]: DEBUG oslo_concurrency.lockutils [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] Acquiring lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.984628] env[62000]: DEBUG oslo_concurrency.lockutils [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] Acquired lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 886.984790] env[62000]: DEBUG nova.network.neutron [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Refreshing network info cache for port e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 886.993158] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 886.993158] env[62000]: value = "task-882477" [ 886.993158] env[62000]: _type = "Task" [ 886.993158] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.003628] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882477, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.054967] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.259566] env[62000]: INFO nova.compute.manager [-] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Took 1.66 seconds to deallocate network for instance. [ 887.259566] env[62000]: DEBUG oslo_vmware.api [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882475, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133779} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.261730] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 887.262066] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 887.262203] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 887.262364] env[62000]: INFO nova.compute.manager [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Took 1.33 seconds to destroy the instance on the hypervisor. [ 887.262600] env[62000]: DEBUG oslo.service.loopingcall [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 887.265750] env[62000]: DEBUG nova.compute.manager [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 887.268786] env[62000]: DEBUG nova.network.neutron [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 887.285919] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.315454] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882469, 'name': CloneVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.392851] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882476, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.487560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.457s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.487839] env[62000]: DEBUG nova.compute.manager [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62000) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 887.491050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 10.966s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.505274] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882477, 'name': CreateVM_Task, 'duration_secs': 0.319423} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.505415] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 887.506276] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.506569] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.506911] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 887.507196] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-34ea8f7f-69e1-4f97-bb42-3205ba3b74da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.513429] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 887.513429] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc10aa-eb5e-13db-13ee-f4c117d330ca" [ 887.513429] env[62000]: _type = "Task" [ 887.513429] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.521550] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc10aa-eb5e-13db-13ee-f4c117d330ca, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.538089] env[62000]: DEBUG nova.network.neutron [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 887.733266] env[62000]: DEBUG nova.network.neutron [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.772483] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.814765] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882469, 'name': CloneVM_Task, 'duration_secs': 1.583851} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.815125] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Created linked-clone VM from snapshot [ 887.816022] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d59c58df-23f1-47d6-9224-d2b2ab4f286d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.824191] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Uploading image fec7d673-ed56-4ee2-8361-5001b5b340d5 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 887.855361] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.855361] env[62000]: value = "vm-201592" [ 887.855361] env[62000]: _type = "VirtualMachine" [ 887.855361] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.855589] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-df86a9a2-335e-4e8a-a26e-657ff77add7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.864769] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease: (returnval){ [ 887.864769] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520fbfc9-0305-45bc-2ac9-4ddbd41c6c66" [ 887.864769] env[62000]: _type = "HttpNfcLease" [ 887.864769] env[62000]: } obtained for exporting VM: (result){ [ 887.864769] env[62000]: value = "vm-201592" [ 887.864769] env[62000]: _type = "VirtualMachine" [ 887.864769] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.865159] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the lease: (returnval){ [ 887.865159] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520fbfc9-0305-45bc-2ac9-4ddbd41c6c66" [ 887.865159] env[62000]: _type = "HttpNfcLease" [ 887.865159] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.866127] env[62000]: DEBUG nova.objects.instance [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.873638] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.873638] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520fbfc9-0305-45bc-2ac9-4ddbd41c6c66" [ 887.873638] env[62000]: _type = "HttpNfcLease" [ 887.873638] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 887.895620] env[62000]: DEBUG oslo_vmware.api [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882476, 'name': PowerOnVM_Task, 'duration_secs': 0.666852} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.895620] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 887.896226] env[62000]: INFO nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Took 9.43 seconds to spawn the instance on the hypervisor. [ 887.896226] env[62000]: DEBUG nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.896938] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d28ff6b-5c0e-4166-a842-0b97318236d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.025960] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc10aa-eb5e-13db-13ee-f4c117d330ca, 'name': SearchDatastore_Task, 'duration_secs': 0.010764} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.028927] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.029253] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 888.029548] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.029736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.029994] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 888.030709] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8e889472-6aef-411a-910d-4cf2f45c90ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.039444] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 888.039673] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 888.040504] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dbf39b9-d73c-4a2a-98d0-6f11f1c90903 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.045998] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 888.045998] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527072cd-918c-bb45-62b8-ec463e45b72a" [ 888.045998] env[62000]: _type = "Task" [ 888.045998] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.053646] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527072cd-918c-bb45-62b8-ec463e45b72a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.061440] env[62000]: DEBUG nova.network.neutron [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.069210] env[62000]: INFO nova.scheduler.client.report [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted allocation for migration 430688a3-72d0-4b3b-a5ec-98b8352758e1 [ 888.142451] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.239987] env[62000]: DEBUG oslo_concurrency.lockutils [req-9b85d7b6-6a20-47ba-a842-02c8608277c1 req-90388f95-4b98-4303-a63e-f2f62688fb67 service nova] Releasing lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.240414] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.240604] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 888.378262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-366c6d8d-44a1-4428-8faf-896f50b9e482 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.907s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.379650] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.238s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.379872] env[62000]: DEBUG nova.compute.manager [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.382374] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26beaa94-6d88-44bf-885c-8768eda37587 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.385287] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.385287] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520fbfc9-0305-45bc-2ac9-4ddbd41c6c66" [ 888.385287] env[62000]: _type = "HttpNfcLease" [ 888.385287] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 888.385826] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 888.385826] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520fbfc9-0305-45bc-2ac9-4ddbd41c6c66" [ 888.385826] env[62000]: _type = "HttpNfcLease" [ 888.385826] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 888.386886] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f047a3e-fa0e-4d1f-bdd4-99a0fab0695a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.392029] env[62000]: DEBUG nova.compute.manager [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 888.392534] env[62000]: DEBUG nova.objects.instance [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.401119] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 888.401457] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 888.513668] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Applying migration context for instance 0a3be3e8-b079-4006-8a46-9b9dd02baa5b as it has an incoming, in-progress migration 4f39862a-db3a-43d6-9aa3-457c8f47a8c9. Migration status is confirming {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 888.515545] env[62000]: INFO nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating resource usage from migration 4f39862a-db3a-43d6-9aa3-457c8f47a8c9 [ 888.518785] env[62000]: INFO nova.compute.manager [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Took 24.35 seconds to build instance. [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 66edb906-f936-4f57-833d-224f36af109e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 8a11689f-fc00-43f8-9215-8d81daa84400 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.552604] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 4e3f804c-f4a2-44ab-bc84-3c97d5803fab is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 3780b22d-c360-4433-9f6b-9d5d1f14b525 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 44a54bd7-a35c-49ed-85ed-346830cee6ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.552604] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553016] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Migration 4f39862a-db3a-43d6-9aa3-457c8f47a8c9 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 888.553016] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 0a3be3e8-b079-4006-8a46-9b9dd02baa5b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553100] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance edc7c99e-d53c-4e0b-991d-c4b693544cdd actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553202] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 9097dd44-563c-460b-b718-cbc1d5eb4a28 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553315] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 72a5bf70-dc6e-4887-abb8-8fbad64bb065 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553432] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance d2f87d82-af5e-4b17-959e-40865a8852b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.553548] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance c354ba41-0915-44cc-9857-42705a56c00f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 888.562262] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527072cd-918c-bb45-62b8-ec463e45b72a, 'name': SearchDatastore_Task, 'duration_secs': 0.009363} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.563112] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5288811e-106a-4731-9811-d931dbd72a54 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.568780] env[62000]: INFO nova.compute.manager [-] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Took 1.30 seconds to deallocate network for instance. [ 888.571440] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b78f9fff-d0d2-4f83-8306-2915d26c3f28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.581240] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9333bff2-a26e-4fbb-9566-0263a81fc139 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 15.999s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.582787] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 888.582787] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c2e445-ca9c-d6da-5ec6-b17be71829d3" [ 888.582787] env[62000]: _type = "Task" [ 888.582787] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.597416] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c2e445-ca9c-d6da-5ec6-b17be71829d3, 'name': SearchDatastore_Task, 'duration_secs': 0.010762} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.597416] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.597416] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 888.597699] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-342eb823-7bf5-425e-8a63-d37277b309c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.606119] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 888.606119] env[62000]: value = "task-882480" [ 888.606119] env[62000]: _type = "Task" [ 888.606119] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.618573] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882480, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.772887] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.864320] env[62000]: DEBUG nova.objects.instance [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.900762] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 888.901103] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-80d1e561-394e-427d-a20d-df0fd25ff3df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.912223] env[62000]: DEBUG oslo_vmware.api [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 888.912223] env[62000]: value = "task-882481" [ 888.912223] env[62000]: _type = "Task" [ 888.912223] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.924293] env[62000]: DEBUG oslo_vmware.api [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882481, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.991590] env[62000]: DEBUG nova.network.neutron [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Updating instance_info_cache with network_info: [{"id": "e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425", "address": "fa:16:3e:3a:a9:d8", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48f9f01-2f", "ovs_interfaceid": "e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.022541] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c393e13c-d64f-4e04-84d9-a06e28306fb2 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.861s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.056339] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.084445] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.121734] env[62000]: DEBUG nova.compute.manager [req-bbd14ed5-b120-4fed-be47-2f9206d1a492 req-d61e8306-0f65-4af6-9fe9-877eb3fffccc service nova] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Received event network-vif-deleted-f3b44c3b-72c6-43c5-b3a6-30559e284b35 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 889.123561] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882480, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493028} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.123743] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 889.123892] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 889.124582] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-42e6e003-293f-4851-9b33-b8015f95f114 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.135388] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 889.135388] env[62000]: value = "task-882482" [ 889.135388] env[62000]: _type = "Task" [ 889.135388] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.145826] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882482, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.373069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.373069] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.373294] env[62000]: DEBUG nova.network.neutron [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.373550] env[62000]: DEBUG nova.objects.instance [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'info_cache' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.421283] env[62000]: DEBUG oslo_vmware.api [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882481, 'name': PowerOffVM_Task, 'duration_secs': 0.300051} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.421643] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 889.421872] env[62000]: DEBUG nova.compute.manager [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.422717] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf3bece-5b76-43f7-8e17-5b2ab9622f75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.498307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "refresh_cache-c354ba41-0915-44cc-9857-42705a56c00f" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.498307] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance network_info: |[{"id": "e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425", "address": "fa:16:3e:3a:a9:d8", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape48f9f01-2f", "ovs_interfaceid": "e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.498307] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3a:a9:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.506650] env[62000]: DEBUG oslo.service.loopingcall [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.506650] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 889.506650] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-694825ab-cddc-4959-8314-110b39c5bd79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.536211] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.536211] env[62000]: value = "task-882483" [ 889.536211] env[62000]: _type = "Task" [ 889.536211] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.544907] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882483, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.559940] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 889.560632] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 889.560632] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3136MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 889.646037] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882482, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07571} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.646783] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 889.647852] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe9e383e-94dc-4aa6-a20f-10c362dc5c94 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.678578] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.681894] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2309b3b7-5399-4d55-b18d-a38957e24bc8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.702638] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.702990] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.710068] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 889.710068] env[62000]: value = "task-882484" [ 889.710068] env[62000]: _type = "Task" [ 889.710068] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.722366] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882484, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.817765] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 889.819334] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201591', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'name': 'volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'edc7c99e-d53c-4e0b-991d-c4b693544cdd', 'attached_at': '', 'detached_at': '', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'serial': '3e32054b-1f06-4121-8f0c-31b7d7ea244a'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 889.819334] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a35c89-fb38-4891-948e-1eb6d4460471 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.838475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4151135b-a422-4e7b-9164-9c68f779ed92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.869158] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a/volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 889.873911] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67c0aee2-689f-4b84-a8eb-b28d6280624c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.891616] env[62000]: DEBUG nova.objects.base [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Object Instance<5a8cded8-bcfb-4488-a736-fb6b6aad5a94> lazy-loaded attributes: flavor,info_cache {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 889.898563] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 889.898563] env[62000]: value = "task-882485" [ 889.898563] env[62000]: _type = "Task" [ 889.898563] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.910346] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882485, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.912306] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66fa218e-63e0-4afd-af72-975440a53653 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.921030] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb935eb6-539a-4f47-a977-95b6011fa78e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.958473] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e88b4a1-f59e-4b07-9fd9-75d94dee0c7a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.579s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.960017] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e9c6db-f093-472a-b608-10d1141fb4c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.968682] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-525f57b9-4feb-46ec-b031-16d908a0cbd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.984603] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 890.048237] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882483, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.206312] env[62000]: DEBUG nova.compute.utils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.221225] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882484, 'name': ReconfigVM_Task, 'duration_secs': 0.399208} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.221537] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to attach disk [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.222933] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-135c7e6c-91ef-41f3-8886-599d5898c764 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.231479] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 890.231479] env[62000]: value = "task-882486" [ 890.231479] env[62000]: _type = "Task" [ 890.231479] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.242261] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882486, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.410764] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882485, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.488482] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 890.552554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.552936] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.567617] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882483, 'name': CreateVM_Task, 'duration_secs': 0.987095} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.569086] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 890.570071] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.570400] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.570876] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 890.571437] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-287cfc64-76cc-459f-b71f-a51d9bec2041 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.577598] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 890.577598] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5224bbb0-610b-f934-9568-225a068740d4" [ 890.577598] env[62000]: _type = "Task" [ 890.577598] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.587715] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5224bbb0-610b-f934-9568-225a068740d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.668759] env[62000]: DEBUG nova.network.neutron [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [{"id": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "address": "fa:16:3e:7f:00:f0", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa40f58be-07", "ovs_interfaceid": "a40f58be-07f1-43ce-afd9-e0ea0c78b634", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.710796] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.741518] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882486, 'name': Rename_Task, 'duration_secs': 0.207988} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.741817] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 890.742120] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-44af44ea-ce83-4774-bec5-5245739a32e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.749031] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 890.749031] env[62000]: value = "task-882487" [ 890.749031] env[62000]: _type = "Task" [ 890.749031] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.765327] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882487, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.908757] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882485, 'name': ReconfigVM_Task, 'duration_secs': 0.621552} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.909122] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfigured VM instance instance-0000004b to attach disk [datastore1] volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a/volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 890.913935] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-94d4b072-cf11-4387-b57c-43a17f7e0f78 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.929652] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 890.929652] env[62000]: value = "task-882488" [ 890.929652] env[62000]: _type = "Task" [ 890.929652] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.938951] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882488, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.999289] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 890.999565] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.509s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.000097] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.132s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.001878] env[62000]: INFO nova.compute.claims [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 891.025892] env[62000]: DEBUG nova.objects.instance [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.068220] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.088462] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5224bbb0-610b-f934-9568-225a068740d4, 'name': SearchDatastore_Task, 'duration_secs': 0.039912} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.088839] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.089042] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 891.089295] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.089448] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.089631] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 891.089898] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-96d4dc68-113c-49ae-8d6c-9d9303b04ce5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.099222] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 891.099545] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 891.100267] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2b74472a-5f85-4f3a-98fb-0c207f41c53d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.105899] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 891.105899] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f64b3a-d8be-730d-9f60-cff69643cbd5" [ 891.105899] env[62000]: _type = "Task" [ 891.105899] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.115361] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f64b3a-d8be-730d-9f60-cff69643cbd5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.172730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-5a8cded8-bcfb-4488-a736-fb6b6aad5a94" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.262126] env[62000]: DEBUG oslo_vmware.api [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882487, 'name': PowerOnVM_Task, 'duration_secs': 0.493012} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.263251] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 891.263251] env[62000]: INFO nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Took 8.52 seconds to spawn the instance on the hypervisor. [ 891.263418] env[62000]: DEBUG nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.266527] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d21f799-b974-4899-afed-d77f915a3b51 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.284079] env[62000]: DEBUG nova.compute.manager [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 891.284296] env[62000]: DEBUG nova.compute.manager [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 891.284540] env[62000]: DEBUG oslo_concurrency.lockutils [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.284718] env[62000]: DEBUG oslo_concurrency.lockutils [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.284919] env[62000]: DEBUG nova.network.neutron [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 891.440833] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882488, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.530779] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.530779] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.531032] env[62000]: DEBUG nova.network.neutron [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.531032] env[62000]: DEBUG nova.objects.instance [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'info_cache' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 891.593389] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.620436] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f64b3a-d8be-730d-9f60-cff69643cbd5, 'name': SearchDatastore_Task, 'duration_secs': 0.010088} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.621616] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b29b8763-43e2-4356-a4f7-ee0d4cf8e819 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.628736] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 891.628736] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249deed-fff1-aaa0-5a15-e41603b6d995" [ 891.628736] env[62000]: _type = "Task" [ 891.628736] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.639639] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249deed-fff1-aaa0-5a15-e41603b6d995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.676290] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 891.677170] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8d4deba7-475f-40c1-85fa-28d8ab44934c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.685291] env[62000]: DEBUG oslo_vmware.api [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 891.685291] env[62000]: value = "task-882489" [ 891.685291] env[62000]: _type = "Task" [ 891.685291] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.694135] env[62000]: DEBUG oslo_vmware.api [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882489, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.771757] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.771951] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.772368] env[62000]: INFO nova.compute.manager [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Attaching volume c6426f66-679d-4eb9-9518-08c638760409 to /dev/sdb [ 891.789921] env[62000]: INFO nova.compute.manager [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Took 20.11 seconds to build instance. [ 891.813070] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-318d0043-52a9-4cb6-bbf3-6ba7ba66bddf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.821384] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f7fe562-82d7-439a-9d06-78e35160f4ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.835245] env[62000]: DEBUG nova.virt.block_device [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating existing volume attachment record: f3fc4f5d-ca85-4d20-a0bd-d90d949f8b02 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 891.941538] env[62000]: DEBUG oslo_vmware.api [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882488, 'name': ReconfigVM_Task, 'duration_secs': 0.749813} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.941953] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201591', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'name': 'volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'edc7c99e-d53c-4e0b-991d-c4b693544cdd', 'attached_at': '', 'detached_at': '', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'serial': '3e32054b-1f06-4121-8f0c-31b7d7ea244a'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 892.035030] env[62000]: DEBUG nova.objects.base [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Object Instance<66edb906-f936-4f57-833d-224f36af109e> lazy-loaded attributes: flavor,info_cache {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 892.146868] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249deed-fff1-aaa0-5a15-e41603b6d995, 'name': SearchDatastore_Task, 'duration_secs': 0.023888} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.147228] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.147522] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c354ba41-0915-44cc-9857-42705a56c00f/c354ba41-0915-44cc-9857-42705a56c00f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 892.147799] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6510b9f2-b4ce-493a-8855-98c11e1dd1b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.155717] env[62000]: DEBUG nova.network.neutron [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 892.155995] env[62000]: DEBUG nova.network.neutron [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.163034] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 892.163034] env[62000]: value = "task-882493" [ 892.163034] env[62000]: _type = "Task" [ 892.163034] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.172857] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882493, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.199367] env[62000]: DEBUG oslo_vmware.api [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882489, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.291540] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ddc8b0dd-a2d0-4ec7-8f45-face7b01d8e0 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.631s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.310841] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9807d14e-cb69-48fd-8812-734ff838de0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.324074] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a43bd5b-4586-4aaa-ba93-761aba6ac816 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.359427] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8638bdef-d5c5-4204-bd81-de54b6219ad6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.367344] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b72be8-3109-4cb5-89b7-a71e5417e572 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.381078] env[62000]: DEBUG nova.compute.provider_tree [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.659862] env[62000]: DEBUG oslo_concurrency.lockutils [req-65664b3a-7cdb-40bc-8784-73674f0c779b req-34cec779-43c8-4fdc-b678-8b32d58b8062 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.675921] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882493, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.701468] env[62000]: DEBUG oslo_vmware.api [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882489, 'name': PowerOnVM_Task, 'duration_secs': 0.604432} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.701468] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 892.701950] env[62000]: DEBUG nova.compute.manager [None req-cdae1892-f548-4699-b6ca-6a86ca15877a tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 892.703115] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1543b52-a541-44bc-9919-daa66729515d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.883932] env[62000]: DEBUG nova.scheduler.client.report [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 893.005364] env[62000]: DEBUG nova.objects.instance [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'flavor' on Instance uuid edc7c99e-d53c-4e0b-991d-c4b693544cdd {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.027280] env[62000]: DEBUG nova.network.neutron [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.162674] env[62000]: DEBUG nova.compute.manager [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Received event network-changed-919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 893.162878] env[62000]: DEBUG nova.compute.manager [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Refreshing instance network info cache due to event network-changed-919b3f5d-af5d-45e1-896f-ac0ea1309017. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 893.163218] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] Acquiring lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.163983] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] Acquired lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.164216] env[62000]: DEBUG nova.network.neutron [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Refreshing network info cache for port 919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 893.176937] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882493, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.579863} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.177209] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c354ba41-0915-44cc-9857-42705a56c00f/c354ba41-0915-44cc-9857-42705a56c00f.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 893.177450] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 893.177668] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b15e4f2-205d-401e-a81a-e8714a536d0a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.184931] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 893.184931] env[62000]: value = "task-882494" [ 893.184931] env[62000]: _type = "Task" [ 893.184931] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.193926] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882494, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.389724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.390864] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.393059] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.362s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.395027] env[62000]: INFO nova.compute.claims [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.511648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62283b84-34a4-423b-971d-7bd9c4bfb639 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.390s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.531701] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.629358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.630144] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.630301] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.630492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.631255] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.633184] env[62000]: INFO nova.compute.manager [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Terminating instance [ 893.635228] env[62000]: DEBUG nova.compute.manager [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 893.635449] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 893.635691] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1c3beeb8-5cb4-4891-9eff-ce4fbeb44101 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.643730] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 893.643730] env[62000]: value = "task-882495" [ 893.643730] env[62000]: _type = "Task" [ 893.643730] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.652432] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882495, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.698763] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882494, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.085599} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.701589] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 893.702410] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f50104-6ba0-4cb5-b912-336dd981dd2e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.724674] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] c354ba41-0915-44cc-9857-42705a56c00f/c354ba41-0915-44cc-9857-42705a56c00f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.725587] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47803050-1135-404e-b445-e9ef095f83fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.752086] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 893.752086] env[62000]: value = "task-882496" [ 893.752086] env[62000]: _type = "Task" [ 893.752086] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.761069] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882496, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.900159] env[62000]: DEBUG nova.compute.utils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.904464] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 893.951685] env[62000]: DEBUG nova.network.neutron [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updated VIF entry in instance network info cache for port 919b3f5d-af5d-45e1-896f-ac0ea1309017. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 893.952243] env[62000]: DEBUG nova.network.neutron [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating instance_info_cache with network_info: [{"id": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "address": "fa:16:3e:86:36:44", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.230", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919b3f5d-af", "ovs_interfaceid": "919b3f5d-af5d-45e1-896f-ac0ea1309017", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.035097] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.035547] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-42d115d1-c83f-4370-82e2-cd7e9726606b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.042787] env[62000]: DEBUG oslo_vmware.api [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 894.042787] env[62000]: value = "task-882497" [ 894.042787] env[62000]: _type = "Task" [ 894.042787] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.052953] env[62000]: DEBUG oslo_vmware.api [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882497, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.154730] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882495, 'name': PowerOffVM_Task, 'duration_secs': 0.315355} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.154910] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.155705] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 894.155705] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201591', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'name': 'volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'edc7c99e-d53c-4e0b-991d-c4b693544cdd', 'attached_at': '', 'detached_at': '', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'serial': '3e32054b-1f06-4121-8f0c-31b7d7ea244a'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 894.156810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80068d59-d28a-43fb-9a32-e1de567d3876 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.179596] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4160aa89-ff21-4c7c-9145-c9e0dc521196 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.186820] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4ead0c-1006-4bba-940e-8a97b91f9b1d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.206669] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.206937] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.207171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.207361] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 894.207535] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 894.209828] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d78fc73-36c7-47b7-a4e7-2512b8f6c617 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.215173] env[62000]: INFO nova.compute.manager [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Terminating instance [ 894.218169] env[62000]: DEBUG nova.compute.manager [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 894.218377] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 894.219188] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc1350de-ab9b-423b-872e-72c78dbc6dbc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.233431] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] The volume has not been displaced from its original location: [datastore1] volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a/volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 894.239033] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfiguring VM instance instance-0000004b to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 894.239613] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cc27e707-9dac-4d4e-9049-67d6fd5bd7eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.255191] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 894.259059] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f45bf97-ada5-4b09-9383-daf196541c71 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.266613] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882496, 'name': ReconfigVM_Task, 'duration_secs': 0.290785} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.269104] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Reconfigured VM instance instance-0000004f to attach disk [datastore1] c354ba41-0915-44cc-9857-42705a56c00f/c354ba41-0915-44cc-9857-42705a56c00f.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.269801] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 894.269801] env[62000]: value = "task-882499" [ 894.269801] env[62000]: _type = "Task" [ 894.269801] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.270095] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 894.270095] env[62000]: value = "task-882500" [ 894.270095] env[62000]: _type = "Task" [ 894.270095] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.270345] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-405aa1d6-d54f-4a6b-b352-1952230e82b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.284169] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882499, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.288274] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.288620] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 894.288620] env[62000]: value = "task-882501" [ 894.288620] env[62000]: _type = "Task" [ 894.288620] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.298499] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882501, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.405887] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.455692] env[62000]: DEBUG oslo_concurrency.lockutils [req-ef3cf01b-5282-476d-8f9c-1ae7b8e24d2c req-58bdf840-0b9c-4f30-be3d-e9ad3202716c service nova] Releasing lock "refresh_cache-d2f87d82-af5e-4b17-959e-40865a8852b3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.555977] env[62000]: DEBUG oslo_vmware.api [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882497, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.672911] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36951dd2-0413-4a2c-86cf-60277aecfba8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.681050] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-346f5945-c513-420b-a045-85c48b41a5c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.713825] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7e82fb-521c-4a7f-918b-a92c017c5d49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.721926] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34cb271-83c3-46ce-adf4-9932d9bedb7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.737572] env[62000]: DEBUG nova.compute.provider_tree [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.787106] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882499, 'name': ReconfigVM_Task, 'duration_secs': 0.397137} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.787372] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882500, 'name': PowerOffVM_Task, 'duration_secs': 0.207091} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.787618] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Reconfigured VM instance instance-0000004b to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 894.792304] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 894.792502] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 894.792796] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a73e5eae-2891-4158-a099-611a19334f5c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.802420] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d17f4776-1d8b-43cd-a6b9-a3ee3947357d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.811785] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882501, 'name': Rename_Task, 'duration_secs': 0.176556} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.813170] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 894.813568] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 894.813568] env[62000]: value = "task-882503" [ 894.813568] env[62000]: _type = "Task" [ 894.813568] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.813833] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a1c1124b-fe9f-4e27-b254-440f0aef1841 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.823093] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882503, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.824343] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 894.824343] env[62000]: value = "task-882504" [ 894.824343] env[62000]: _type = "Task" [ 894.824343] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.831951] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882504, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.876790] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 894.877084] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 894.877280] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleting the datastore file [datastore1] 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 894.877570] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0694a696-b017-4660-9f0f-67372b851168 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.884535] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 894.884535] env[62000]: value = "task-882505" [ 894.884535] env[62000]: _type = "Task" [ 894.884535] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.892257] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882505, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.055053] env[62000]: DEBUG oslo_vmware.api [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882497, 'name': PowerOnVM_Task, 'duration_secs': 0.548077} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.055053] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 895.055660] env[62000]: DEBUG nova.compute.manager [None req-d4fd33ad-ab84-4be0-a86a-6182e4cf9c4e tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.056040] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acf10066-c310-483f-81d0-656b192c8ab9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.244168] env[62000]: DEBUG nova.scheduler.client.report [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 895.325112] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882503, 'name': ReconfigVM_Task, 'duration_secs': 0.316493} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.328864] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201591', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'name': 'volume-3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'edc7c99e-d53c-4e0b-991d-c4b693544cdd', 'attached_at': '', 'detached_at': '', 'volume_id': '3e32054b-1f06-4121-8f0c-31b7d7ea244a', 'serial': '3e32054b-1f06-4121-8f0c-31b7d7ea244a'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 895.329251] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 895.330663] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a8dc81-0767-4e18-aa77-6479e382c7b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.338753] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882504, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.340994] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 895.341267] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-477fd866-8454-4891-b8bb-6179f547461a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.395697] env[62000]: DEBUG oslo_vmware.api [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882505, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352535} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.396077] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.396288] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.396516] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.396725] env[62000]: INFO nova.compute.manager [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 1.18 seconds to destroy the instance on the hypervisor. [ 895.396991] env[62000]: DEBUG oslo.service.loopingcall [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.397220] env[62000]: DEBUG nova.compute.manager [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.397317] env[62000]: DEBUG nova.network.neutron [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 895.419640] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.422710] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 895.422930] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 895.423242] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore2] edc7c99e-d53c-4e0b-991d-c4b693544cdd {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 895.423579] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1df85c0d-39f5-4a20-a747-97a88f9b195b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.432055] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 895.432055] env[62000]: value = "task-882507" [ 895.432055] env[62000]: _type = "Task" [ 895.432055] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.443591] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882507, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.452534] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.452847] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.453053] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.453275] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.453494] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.453689] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.453942] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.454143] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.454362] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.454553] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.454766] env[62000]: DEBUG nova.virt.hardware [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.455755] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d184bd03-dfe0-4d66-a8a3-b730d6879f8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.464160] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fba772f-d83b-43d5-8b47-ca34ccd4b399 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.478859] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.485162] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Creating folder: Project (97cb28d687674dbf8eb7184eade17524). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.485564] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30b1f40e-de13-459b-ba39-050c8139051c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.497619] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Created folder: Project (97cb28d687674dbf8eb7184eade17524) in parent group-v201431. [ 895.497917] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Creating folder: Instances. Parent ref: group-v201597. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 895.498270] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-239821ea-f820-43d9-9d13-12c18e53f874 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.509099] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Created folder: Instances in parent group-v201597. [ 895.509538] env[62000]: DEBUG oslo.service.loopingcall [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.509922] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 895.510227] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dcd0ef9b-a41b-4cc2-9d01-031bd75655b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.543829] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.543829] env[62000]: value = "task-882510" [ 895.543829] env[62000]: _type = "Task" [ 895.543829] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.553498] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882510, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.749461] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.750178] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.753651] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 13.131s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.837933] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882504, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.942404] env[62000]: DEBUG oslo_vmware.api [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882507, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.350586} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.942772] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 895.943075] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 895.943351] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 895.943562] env[62000]: INFO nova.compute.manager [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Took 2.31 seconds to destroy the instance on the hypervisor. [ 895.943875] env[62000]: DEBUG oslo.service.loopingcall [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.944160] env[62000]: DEBUG nova.compute.manager [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 895.944305] env[62000]: DEBUG nova.network.neutron [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.001651] env[62000]: DEBUG nova.compute.manager [req-12691948-65c0-44be-a1cd-65ca668a731a req-53a03b40-1501-4195-8c11-7a870a4a43e5 service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Received event network-vif-deleted-a40f58be-07f1-43ce-afd9-e0ea0c78b634 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.001890] env[62000]: INFO nova.compute.manager [req-12691948-65c0-44be-a1cd-65ca668a731a req-53a03b40-1501-4195-8c11-7a870a4a43e5 service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Neutron deleted interface a40f58be-07f1-43ce-afd9-e0ea0c78b634; detaching it from the instance and deleting it from the info cache [ 896.002138] env[62000]: DEBUG nova.network.neutron [req-12691948-65c0-44be-a1cd-65ca668a731a req-53a03b40-1501-4195-8c11-7a870a4a43e5 service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.055152] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882510, 'name': CreateVM_Task, 'duration_secs': 0.419312} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.055545] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 896.056317] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.056317] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.056504] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 896.056613] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fce3fa0-06f8-462a-a96c-229ff2b4c22f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.062617] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 896.062617] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529d09bd-551d-0aa0-6536-5553c44f861b" [ 896.062617] env[62000]: _type = "Task" [ 896.062617] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.072108] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529d09bd-551d-0aa0-6536-5553c44f861b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.222334] env[62000]: DEBUG nova.compute.manager [req-bfe87ffc-f6b7-45ca-be38-b0423980ebd1 req-efaef654-0f29-4a15-81cf-4354498aa590 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Received event network-vif-deleted-9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 896.223029] env[62000]: INFO nova.compute.manager [req-bfe87ffc-f6b7-45ca-be38-b0423980ebd1 req-efaef654-0f29-4a15-81cf-4354498aa590 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Neutron deleted interface 9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e; detaching it from the instance and deleting it from the info cache [ 896.223029] env[62000]: DEBUG nova.network.neutron [req-bfe87ffc-f6b7-45ca-be38-b0423980ebd1 req-efaef654-0f29-4a15-81cf-4354498aa590 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.257517] env[62000]: DEBUG nova.compute.utils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 896.262505] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Not allocating networking since 'none' was specified. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 896.334095] env[62000]: DEBUG oslo_vmware.api [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882504, 'name': PowerOnVM_Task, 'duration_secs': 1.352252} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.334384] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 896.334594] env[62000]: INFO nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Took 11.19 seconds to spawn the instance on the hypervisor. [ 896.334776] env[62000]: DEBUG nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.335629] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a578d44-b3d4-472d-b053-6bdee0952054 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.382975] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 896.383392] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201596', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'name': 'volume-c6426f66-679d-4eb9-9518-08c638760409', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3780b22d-c360-4433-9f6b-9d5d1f14b525', 'attached_at': '', 'detached_at': '', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'serial': 'c6426f66-679d-4eb9-9518-08c638760409'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 896.384341] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c24247-19a1-498b-b67c-8754aa630eaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.407073] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7752986-f941-406a-b211-f6e74f4f8113 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.433110] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] volume-c6426f66-679d-4eb9-9518-08c638760409/volume-c6426f66-679d-4eb9-9518-08c638760409.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 896.435930] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fe05969-e27c-4ee9-9a3f-e1ae2a40e94f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.455891] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 896.455891] env[62000]: value = "task-882511" [ 896.455891] env[62000]: _type = "Task" [ 896.455891] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.466032] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882511, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.478633] env[62000]: DEBUG nova.network.neutron [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.504672] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0e53487e-7fad-43f5-bd74-c1144a9d232a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.516861] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab6502c-2906-4017-a7ef-df9df6ec7baa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.553470] env[62000]: DEBUG nova.compute.manager [req-12691948-65c0-44be-a1cd-65ca668a731a req-53a03b40-1501-4195-8c11-7a870a4a43e5 service nova] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Detach interface failed, port_id=a40f58be-07f1-43ce-afd9-e0ea0c78b634, reason: Instance 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 896.574704] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529d09bd-551d-0aa0-6536-5553c44f861b, 'name': SearchDatastore_Task, 'duration_secs': 0.017829} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.575031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.575276] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.575505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.575650] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.575824] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.576118] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e281c6c-67d6-4c23-8d1f-9a99d265a8f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.579765] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62dd4e78-1622-4875-acf9-1b6c9790db86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.588053] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cb67ee3-19f9-4e2f-8dff-214a0e836aa6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.591504] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.591688] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 896.592452] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd8016e0-64f4-4230-8544-3ab7adfc849a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.597790] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 896.597790] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5274f6ae-afd0-3632-b883-c2bc940d08ba" [ 896.597790] env[62000]: _type = "Task" [ 896.597790] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.626964] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d9b91d-b504-4b4c-8e12-f1dea9152084 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.634756] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5274f6ae-afd0-3632-b883-c2bc940d08ba, 'name': SearchDatastore_Task, 'duration_secs': 0.013578} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.637513] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cfb00419-09fb-4c54-ac47-552f9d493e37 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.640624] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e67f03-4fa7-49f4-acdf-948ffb36e1bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.655043] env[62000]: DEBUG nova.compute.provider_tree [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.656746] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 896.656746] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf6b4c-0a03-aa99-a07a-12337f1aff2c" [ 896.656746] env[62000]: _type = "Task" [ 896.656746] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.664858] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf6b4c-0a03-aa99-a07a-12337f1aff2c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.684818] env[62000]: DEBUG nova.network.neutron [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.728029] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a410c80b-7c00-4087-9bb5-876d5b74bbf3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.737322] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e0aee78-9c5a-485d-bc6b-09ad858eb388 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.770745] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 896.774314] env[62000]: DEBUG nova.compute.manager [req-bfe87ffc-f6b7-45ca-be38-b0423980ebd1 req-efaef654-0f29-4a15-81cf-4354498aa590 service nova] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Detach interface failed, port_id=9d6e8ee9-bc00-41bc-a4d6-ee91a0649f2e, reason: Instance edc7c99e-d53c-4e0b-991d-c4b693544cdd could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 896.860860] env[62000]: INFO nova.compute.manager [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Took 23.53 seconds to build instance. [ 896.966818] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882511, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.983512] env[62000]: INFO nova.compute.manager [-] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Took 1.59 seconds to deallocate network for instance. [ 897.158225] env[62000]: DEBUG nova.scheduler.client.report [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 897.171537] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cf6b4c-0a03-aa99-a07a-12337f1aff2c, 'name': SearchDatastore_Task, 'duration_secs': 0.019116} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.171837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.172150] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b/0f5b416a-20e4-42f5-9ad9-a8489ab11e3b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 897.172425] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d45883ed-6197-435e-bb67-4f863e0eb501 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.179450] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 897.179450] env[62000]: value = "task-882512" [ 897.179450] env[62000]: _type = "Task" [ 897.179450] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.187695] env[62000]: INFO nova.compute.manager [-] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Took 1.24 seconds to deallocate network for instance. [ 897.187989] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882512, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.363596] env[62000]: DEBUG oslo_concurrency.lockutils [None req-18b8a964-5bac-4194-a545-da8e7dcdf4d8 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.046s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.467105] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882511, 'name': ReconfigVM_Task, 'duration_secs': 0.59287} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.467418] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfigured VM instance instance-00000047 to attach disk [datastore1] volume-c6426f66-679d-4eb9-9518-08c638760409/volume-c6426f66-679d-4eb9-9518-08c638760409.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 897.473317] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d0b4697c-d44d-4ccf-89bf-ef5e67f6933d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.488625] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 897.488625] env[62000]: value = "task-882513" [ 897.488625] env[62000]: _type = "Task" [ 897.488625] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.494664] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.500220] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882513, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.693703] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882512, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.736060] env[62000]: INFO nova.compute.manager [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Took 0.55 seconds to detach 1 volumes for instance. [ 897.781115] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 897.811473] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.811841] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.812072] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.812300] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.812529] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.812685] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.813022] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.813254] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.813484] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.813730] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.813965] env[62000]: DEBUG nova.virt.hardware [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.815471] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cf7f1cb-2d08-40cd-818d-4dad57e044ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.824872] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dd61600-1ec8-4571-a742-dd0f58793512 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.842365] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.848429] env[62000]: DEBUG oslo.service.loopingcall [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.849177] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 897.849437] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ab824555-0c2c-4447-99bf-55ad72b0a22f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.868126] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.868126] env[62000]: value = "task-882514" [ 897.868126] env[62000]: _type = "Task" [ 897.868126] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.877562] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882514, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.943964] env[62000]: DEBUG nova.compute.manager [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.944961] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-676d2aa4-eaef-46fe-9891-d256c04b6412 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.001902] env[62000]: DEBUG oslo_vmware.api [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882513, 'name': ReconfigVM_Task, 'duration_secs': 0.178539} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.002618] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201596', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'name': 'volume-c6426f66-679d-4eb9-9518-08c638760409', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3780b22d-c360-4433-9f6b-9d5d1f14b525', 'attached_at': '', 'detached_at': '', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'serial': 'c6426f66-679d-4eb9-9518-08c638760409'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 898.173372] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.420s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.176689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.957s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.176929] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.177202] env[62000]: INFO nova.compute.manager [None req-1eabd9df-4055-4bca-8a74-4b1a4d3ec856 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Successfully reverted task state from rebuilding on failure for instance. [ 898.184360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.129s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.184360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.185951] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.414s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.186166] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.187850] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.104s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.188076] env[62000]: DEBUG nova.objects.instance [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid 9097dd44-563c-460b-b718-cbc1d5eb4a28 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 898.199709] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882512, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.754446} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.200611] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b/0f5b416a-20e4-42f5-9ad9-a8489ab11e3b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 898.200611] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.200611] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cf35fead-7899-4fa4-93db-d39304e42707 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.209511] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 898.209511] env[62000]: value = "task-882515" [ 898.209511] env[62000]: _type = "Task" [ 898.209511] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.214352] env[62000]: INFO nova.scheduler.client.report [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Deleted allocations for instance 4e3f804c-f4a2-44ab-bc84-3c97d5803fab [ 898.218968] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882515, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.242788] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.380454] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882514, 'name': CreateVM_Task, 'duration_secs': 0.36737} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.380684] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 898.380992] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.381182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.381510] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.381831] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4332553f-4755-4aa5-a5ab-78ecc52a6e36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.386204] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 898.386204] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f0825e-40fc-8a05-177a-55eb9b81490a" [ 898.386204] env[62000]: _type = "Task" [ 898.386204] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.393247] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f0825e-40fc-8a05-177a-55eb9b81490a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.456494] env[62000]: INFO nova.compute.manager [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] instance snapshotting [ 898.459908] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-597065dd-76b7-42aa-a29b-dfc6bc9559bb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.479365] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f517d5-44f3-4889-8a35-b7b7d2870a16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.714732] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f281eb49-bd6d-4925-985c-f2b8249db361 tempest-ServerActionsV293TestJSON-2059498064 tempest-ServerActionsV293TestJSON-2059498064-project-member] Lock "2f35dab3-6c32-4a35-91ca-1d6859e27f2c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.990s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.726331] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882515, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076211} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.729055] env[62000]: DEBUG oslo_concurrency.lockutils [None req-38676cf5-413f-46f9-8f20-af5062dc0f0a tempest-ServersTestFqdnHostnames-910858584 tempest-ServersTestFqdnHostnames-910858584-project-member] Lock "4e3f804c-f4a2-44ab-bc84-3c97d5803fab" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.362s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.730051] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.733623] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7ccc2d-1e56-4c90-87fc-640e202436fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.753601] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b/0f5b416a-20e4-42f5-9ad9-a8489ab11e3b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.755339] env[62000]: INFO nova.scheduler.client.report [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocation for migration 4f39862a-db3a-43d6-9aa3-457c8f47a8c9 [ 898.760389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-65b928c9-2700-47a7-887c-41220ff0d002 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.781551] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 898.781551] env[62000]: value = "task-882516" [ 898.781551] env[62000]: _type = "Task" [ 898.781551] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.791240] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882516, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.824890] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 898.826182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8511919-6111-4ef0-9476-afa3338adab8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.833371] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 898.833371] env[62000]: ERROR oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk due to incomplete transfer. [ 898.836029] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c92b1f1f-4769-4d5f-aba9-247af672ce9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.843399] env[62000]: DEBUG oslo_vmware.rw_handles [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/522abe02-0b18-0abb-ac6c-4a4daab9c386/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 898.843637] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Uploaded image fec7d673-ed56-4ee2-8361-5001b5b340d5 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 898.845909] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 898.846898] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-5600dce0-9dbd-404c-a402-12cc5b21d18d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.852620] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 898.852620] env[62000]: value = "task-882517" [ 898.852620] env[62000]: _type = "Task" [ 898.852620] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.860536] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882517, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.902278] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f0825e-40fc-8a05-177a-55eb9b81490a, 'name': SearchDatastore_Task, 'duration_secs': 0.026985} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.902278] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.902278] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.902278] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.902278] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.902278] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.902278] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e46f6856-16e6-496a-97aa-a3defcf6af03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.910914] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.912463] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 898.912463] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-679323bb-3e9c-4efe-ae56-2e816ddd1a6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.917686] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 898.917686] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529c2b9a-3dc4-9f55-d2d6-a9e95aee6321" [ 898.917686] env[62000]: _type = "Task" [ 898.917686] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.927908] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529c2b9a-3dc4-9f55-d2d6-a9e95aee6321, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.990346] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 898.991421] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be818404-9ed3-44a0-a389-60dbbcdbda80 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.994436] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-26ae0050-7d74-4bb5-b765-8e7bb06c0f10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.002056] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375cd539-ed05-436c-b1fa-d97eebae1015 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.006813] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 899.006813] env[62000]: value = "task-882518" [ 899.006813] env[62000]: _type = "Task" [ 899.006813] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.038122] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d6ae6bb-11ec-44ee-ac6b-6da41e241fb5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.044654] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882518, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.046036] env[62000]: DEBUG nova.objects.instance [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 3780b22d-c360-4433-9f6b-9d5d1f14b525 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.052127] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a4ebb2-68a0-4a60-81af-2a2b5019bf77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.064639] env[62000]: DEBUG nova.compute.provider_tree [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.277663] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62dd147b-2e0f-4435-b96e-e5a80da1ee40 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 19.608s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.291694] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882516, 'name': ReconfigVM_Task, 'duration_secs': 0.352754} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.292069] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b/0f5b416a-20e4-42f5-9ad9-a8489ab11e3b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.293398] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e5301c67-3ea7-4da1-9314-ec008b408375 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.301931] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 899.301931] env[62000]: value = "task-882519" [ 899.301931] env[62000]: _type = "Task" [ 899.301931] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.312128] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882519, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.365332] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882517, 'name': Destroy_Task, 'duration_secs': 0.478018} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.368606] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Destroyed the VM [ 899.368606] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 899.368606] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9b124538-b012-4e89-ba4c-af147e76d144 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.372905] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 899.372905] env[62000]: value = "task-882520" [ 899.372905] env[62000]: _type = "Task" [ 899.372905] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.384973] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882520, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.428166] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529c2b9a-3dc4-9f55-d2d6-a9e95aee6321, 'name': SearchDatastore_Task, 'duration_secs': 0.012749} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.429040] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-563765e5-1269-4846-a9c0-c07b864722ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.435164] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 899.435164] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526b1af6-d993-e0b6-8417-62314d18de53" [ 899.435164] env[62000]: _type = "Task" [ 899.435164] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.444694] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526b1af6-d993-e0b6-8417-62314d18de53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.517967] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882518, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.553366] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3031a29b-7614-4805-9052-9473255c0c85 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.781s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.569633] env[62000]: DEBUG nova.scheduler.client.report [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 899.812072] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882519, 'name': Rename_Task, 'duration_secs': 0.170548} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.812433] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 899.812821] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-672ee4f6-c97d-4a4e-9a9f-53675bfd3c4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.819806] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 899.819806] env[62000]: value = "task-882521" [ 899.819806] env[62000]: _type = "Task" [ 899.819806] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.827639] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.882529] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882520, 'name': RemoveSnapshot_Task, 'duration_secs': 0.410662} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.882747] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 899.883192] env[62000]: DEBUG nova.compute.manager [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 899.884078] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b10278-7850-45ee-85e0-51e5870f09b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.946512] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526b1af6-d993-e0b6-8417-62314d18de53, 'name': SearchDatastore_Task, 'duration_secs': 0.021721} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.946837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.947208] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 899.947500] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b40cbb6-ee24-4628-a1e4-d5a4f47d9442 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.956340] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 899.956340] env[62000]: value = "task-882522" [ 899.956340] env[62000]: _type = "Task" [ 899.956340] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.965625] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882522, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.018598] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882518, 'name': CreateSnapshot_Task, 'duration_secs': 1.014286} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.018931] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 900.019731] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4e7da1a-f5bc-4642-a092-08ebe72337f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.076844] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.889s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.083028] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.486s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.083028] env[62000]: INFO nova.compute.claims [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 900.121305] env[62000]: INFO nova.scheduler.client.report [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance 9097dd44-563c-460b-b718-cbc1d5eb4a28 [ 900.335023] env[62000]: DEBUG oslo_vmware.api [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882521, 'name': PowerOnVM_Task, 'duration_secs': 0.477748} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.335023] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 900.335023] env[62000]: INFO nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Took 4.91 seconds to spawn the instance on the hypervisor. [ 900.335023] env[62000]: DEBUG nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.335023] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2722752e-a9f8-4f2f-adfe-c7618cd08b4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.398289] env[62000]: INFO nova.compute.manager [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Shelve offloading [ 900.399225] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 900.399619] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fe760d33-10fe-4928-ba89-edddf9ba276a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.407942] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 900.407942] env[62000]: value = "task-882523" [ 900.407942] env[62000]: _type = "Task" [ 900.407942] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.419895] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 900.420958] env[62000]: DEBUG nova.compute.manager [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.421953] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94ede18b-f3c0-46c0-904b-743d91c2af3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.428398] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.430023] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.430023] env[62000]: DEBUG nova.network.neutron [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 900.466418] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882522, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.539467] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 900.540159] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b2b55292-75d3-4a81-ae49-6c3c7fa15719 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.551737] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 900.551737] env[62000]: value = "task-882524" [ 900.551737] env[62000]: _type = "Task" [ 900.551737] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.564756] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.637056] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5e541ee0-a972-4f6f-ac07-fafeb0df25a7 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "9097dd44-563c-460b-b718-cbc1d5eb4a28" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.713s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.851193] env[62000]: INFO nova.compute.manager [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Took 23.01 seconds to build instance. [ 900.968799] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882522, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718274} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.969086] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 900.969286] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 900.969559] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-980ec7a6-7df6-4243-9246-79b2b7cb9347 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.980664] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 900.980664] env[62000]: value = "task-882525" [ 900.980664] env[62000]: _type = "Task" [ 900.980664] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.987997] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882525, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.061648] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.293897] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.294287] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.354302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5ae2f76-80e4-49ad-83e0-a842cb59446e tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.522s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.368436] env[62000]: DEBUG nova.network.neutron [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.380450] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.380696] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.380904] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.382375] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.382704] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.385348] env[62000]: INFO nova.compute.manager [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Terminating instance [ 901.393933] env[62000]: DEBUG nova.compute.manager [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 901.394215] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 901.395276] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e1ad280-18e6-4062-9f81-ebd316e7c91a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.405633] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 901.408137] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8831ede9-3253-4bdc-93cd-6e68c28a22cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.411423] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddeaafb0-b91c-4515-b801-d86dfe13d6d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.421182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da5cdc18-6bfe-45e5-882a-4f80e5cd8f3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.424778] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 901.424778] env[62000]: value = "task-882526" [ 901.424778] env[62000]: _type = "Task" [ 901.424778] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.454672] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a67ed4c-accc-465f-a2fd-0ecb02a72a08 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.460263] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882526, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.465376] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308dfbf6-5132-40a9-a8cc-10b02b0fdf5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.479016] env[62000]: DEBUG nova.compute.provider_tree [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.488701] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882525, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076583} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.489160] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.490612] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc6c62d-cc61-41dc-820e-24d04891ce80 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.513141] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.513655] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb4ea72b-06f7-4cc9-b906-0619071593de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.533923] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 901.533923] env[62000]: value = "task-882527" [ 901.533923] env[62000]: _type = "Task" [ 901.533923] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.542499] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882527, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.561892] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.800937] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 901.871033] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.934965] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882526, 'name': PowerOffVM_Task, 'duration_secs': 0.353922} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.936527] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 901.936720] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 901.939389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6411e440-868f-4a90-8945-bedffc9d86e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.941392] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.941617] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.986346] env[62000]: DEBUG nova.scheduler.client.report [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.012703] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.013061] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.013460] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore2] 0a3be3e8-b079-4006-8a46-9b9dd02baa5b {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.013671] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1dba53e1-9a2d-4931-8b66-e21390b45594 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.021633] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 902.021633] env[62000]: value = "task-882529" [ 902.021633] env[62000]: _type = "Task" [ 902.021633] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.029865] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882529, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.043783] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882527, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.063240] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.267971] env[62000]: DEBUG nova.compute.manager [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-vif-unplugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 902.268251] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.268464] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.268638] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.268833] env[62000]: DEBUG nova.compute.manager [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] No waiting events found dispatching network-vif-unplugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 902.268985] env[62000]: WARNING nova.compute.manager [req-cc6d50af-f77a-43fe-a930-861b4064b5ee req-470b3543-6f0b-42a4-8e99-164a48eb556f service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received unexpected event network-vif-unplugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 for instance with vm_state shelved and task_state shelving_offloading. [ 902.324497] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.380944] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 902.382053] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d076a89-2b2d-41fc-bd6e-3c3d99acf272 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.390552] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 902.390807] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-94f73993-d4fe-4cda-95dd-04ca4b063203 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.445234] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.456727] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "bf606000-346b-48db-972a-000b54a8ec5d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.457346] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.491331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.412s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.491820] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 902.494992] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.495303] env[62000]: DEBUG nova.objects.instance [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'resources' on Instance uuid 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.508158] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 902.508416] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 902.508716] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.509121] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-83fdea49-ed59-447f-8991-1f439966eb4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.516683] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 902.516683] env[62000]: value = "task-882531" [ 902.516683] env[62000]: _type = "Task" [ 902.516683] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.524495] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882531, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.532558] env[62000]: DEBUG oslo_vmware.api [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882529, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165381} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.532788] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 902.532975] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 902.533178] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 902.533354] env[62000]: INFO nova.compute.manager [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Took 1.14 seconds to destroy the instance on the hypervisor. [ 902.533592] env[62000]: DEBUG oslo.service.loopingcall [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 902.533783] env[62000]: DEBUG nova.compute.manager [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 902.533879] env[62000]: DEBUG nova.network.neutron [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.543118] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882527, 'name': ReconfigVM_Task, 'duration_secs': 0.574828} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.543384] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.543960] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6d2c87de-72a2-49cf-b302-23792ee707e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.551235] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 902.551235] env[62000]: value = "task-882532" [ 902.551235] env[62000]: _type = "Task" [ 902.551235] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.562382] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882532, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.565372] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.959954] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 902.965639] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.998425] env[62000]: DEBUG nova.compute.utils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 902.999914] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 903.000119] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 903.028648] env[62000]: DEBUG oslo_vmware.api [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882531, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13284} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.029663] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.029663] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 903.029663] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 903.041754] env[62000]: DEBUG nova.policy [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.046519] env[62000]: INFO nova.scheduler.client.report [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted allocations for instance 8a11689f-fc00-43f8-9215-8d81daa84400 [ 903.066077] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.068833] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882532, 'name': Rename_Task, 'duration_secs': 0.241454} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.068998] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 903.069231] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d8f6abd-d422-40f3-9bc9-5dd21cfe2b54 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.075592] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 903.075592] env[62000]: value = "task-882533" [ 903.075592] env[62000]: _type = "Task" [ 903.075592] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.083917] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882533, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.233203] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a345878-0569-4c4e-a453-da2b10ef63fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.241215] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e4d75a8-1ff2-400d-b858-79c9604e5e03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.280962] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b876390b-4d96-4391-82ed-7945879b1db4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.288316] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bfb9039-3207-43ec-b6ca-f8ac360fd6e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.295554] env[62000]: DEBUG nova.network.neutron [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.307319] env[62000]: DEBUG nova.compute.provider_tree [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.367466] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Successfully created port: 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 903.481104] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.503543] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 903.550989] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.567542] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.585789] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882533, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.798169] env[62000]: INFO nova.compute.manager [-] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Took 1.26 seconds to deallocate network for instance. [ 903.815022] env[62000]: DEBUG nova.scheduler.client.report [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 904.068954] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task} progress is 95%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.087776] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882533, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.300692] env[62000]: DEBUG nova.compute.manager [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 904.300908] env[62000]: DEBUG nova.compute.manager [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing instance network info cache due to event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 904.301430] env[62000]: DEBUG oslo_concurrency.lockutils [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.301605] env[62000]: DEBUG oslo_concurrency.lockutils [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.301782] env[62000]: DEBUG nova.network.neutron [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 904.309696] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.319859] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.825s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.321792] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.079s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.322064] env[62000]: DEBUG nova.objects.instance [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'resources' on Instance uuid edc7c99e-d53c-4e0b-991d-c4b693544cdd {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 904.335785] env[62000]: INFO nova.scheduler.client.report [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted allocations for instance 5a8cded8-bcfb-4488-a736-fb6b6aad5a94 [ 904.513644] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 904.542220] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 904.542514] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 904.542670] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 904.542853] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 904.543010] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 904.543173] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 904.543474] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 904.543568] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 904.543706] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 904.543927] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 904.544060] env[62000]: DEBUG nova.virt.hardware [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 904.545349] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9ca7a7-0070-46be-be76-c78b32854eec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.553944] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-012bba33-22d9-48d0-a746-d4d745a39b8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.574697] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882524, 'name': CloneVM_Task, 'duration_secs': 3.967384} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.574980] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Created linked-clone VM from snapshot [ 904.575729] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6b4f58c-b196-4f39-bdc2-b322bab4049c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.585910] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Uploading image d19f59f0-69e8-45e4-9559-cc7320fda017 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 904.590440] env[62000]: DEBUG oslo_vmware.api [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882533, 'name': PowerOnVM_Task, 'duration_secs': 1.074548} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.590735] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 904.591093] env[62000]: INFO nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Took 6.81 seconds to spawn the instance on the hypervisor. [ 904.591093] env[62000]: DEBUG nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.591790] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c9bc4e-e81c-4ffa-a606-fb3780ba6739 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.608485] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 904.608485] env[62000]: value = "vm-201602" [ 904.608485] env[62000]: _type = "VirtualMachine" [ 904.608485] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 904.608665] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-15535806-c639-4b49-bbb6-3066aef135f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.615358] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease: (returnval){ [ 904.615358] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52705c6a-f87d-8236-4146-001c9f63c027" [ 904.615358] env[62000]: _type = "HttpNfcLease" [ 904.615358] env[62000]: } obtained for exporting VM: (result){ [ 904.615358] env[62000]: value = "vm-201602" [ 904.615358] env[62000]: _type = "VirtualMachine" [ 904.615358] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 904.615632] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the lease: (returnval){ [ 904.615632] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52705c6a-f87d-8236-4146-001c9f63c027" [ 904.615632] env[62000]: _type = "HttpNfcLease" [ 904.615632] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 904.621684] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 904.621684] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52705c6a-f87d-8236-4146-001c9f63c027" [ 904.621684] env[62000]: _type = "HttpNfcLease" [ 904.621684] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 904.844607] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ee929d98-5258-4c48-a630-76eb79087110 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "5a8cded8-bcfb-4488-a736-fb6b6aad5a94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.638s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.957247] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Successfully updated port: 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 905.028727] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.059680] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647f1ee4-ca84-4f98-b7bc-41a921abe596 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.067658] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-461d3d1d-ec87-456e-965b-ef760edf7c03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.101836] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-173bc9cf-9c83-42eb-9049-ff538ffb3f28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.118928] env[62000]: INFO nova.compute.manager [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Took 27.10 seconds to build instance. [ 905.123782] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d839c461-64c3-496d-9e3b-62aa2f4bf4e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.137657] env[62000]: DEBUG nova.compute.provider_tree [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.140782] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 905.140782] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52705c6a-f87d-8236-4146-001c9f63c027" [ 905.140782] env[62000]: _type = "HttpNfcLease" [ 905.140782] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 905.141251] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 905.141251] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52705c6a-f87d-8236-4146-001c9f63c027" [ 905.141251] env[62000]: _type = "HttpNfcLease" [ 905.141251] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 905.141948] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af1962d-05d4-4a8e-a046-1be9f4664aab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.150232] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 905.150353] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 905.242144] env[62000]: DEBUG nova.network.neutron [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updated VIF entry in instance network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 905.242517] env[62000]: DEBUG nova.network.neutron [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.243793] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-da9f3cb1-1c79-434e-bdfa-794c47b1bea8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.306623] env[62000]: INFO nova.compute.manager [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Rebuilding instance [ 905.354754] env[62000]: DEBUG nova.compute.manager [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.355631] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cc64d82-56d0-42b7-a8d5-26014ae6a246 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.460064] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.460236] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.460390] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 905.628656] env[62000]: DEBUG oslo_concurrency.lockutils [None req-93fa616e-7e3d-434c-b38c-05ac9b3995bf tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.617s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.642973] env[62000]: DEBUG nova.scheduler.client.report [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 905.749141] env[62000]: DEBUG oslo_concurrency.lockutils [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.749749] env[62000]: DEBUG nova.compute.manager [req-ddec3b79-152a-4f03-aee6-582bfe7f0e4e req-9f0bb028-39bb-4bae-b0b3-ee4ac1f3a1a4 service nova] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Received event network-vif-deleted-7e199318-132a-4975-92a1-61d761a06165 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 905.866608] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 905.867117] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-612d3c7f-143b-4da5-b5c5-feebbb849276 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.876356] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 905.876356] env[62000]: value = "task-882535" [ 905.876356] env[62000]: _type = "Task" [ 905.876356] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.886178] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882535, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.922077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.923042] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.993707] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 906.149019] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.151724] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.827s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.153690] env[62000]: INFO nova.compute.claims [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.173417] env[62000]: INFO nova.scheduler.client.report [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance edc7c99e-d53c-4e0b-991d-c4b693544cdd [ 906.197228] env[62000]: DEBUG nova.network.neutron [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.329163] env[62000]: DEBUG nova.compute.manager [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-vif-plugged-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.329512] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.329805] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.330166] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.330391] env[62000]: DEBUG nova.compute.manager [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] No waiting events found dispatching network-vif-plugged-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 906.330594] env[62000]: WARNING nova.compute.manager [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received unexpected event network-vif-plugged-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 for instance with vm_state building and task_state spawning. [ 906.330960] env[62000]: DEBUG nova.compute.manager [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 906.331231] env[62000]: DEBUG nova.compute.manager [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing instance network info cache due to event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 906.331526] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.386458] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882535, 'name': PowerOffVM_Task, 'duration_secs': 0.150125} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.386824] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 906.387210] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 906.388155] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88069410-267e-4142-98a0-5d5a8edbdbb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.395238] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 906.395519] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bbe37463-a950-4c2b-a251-bf5ada7ec422 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.421703] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 906.422083] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 906.422382] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleting the datastore file [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.422700] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a84bd4bd-ec26-41c7-8dd5-e6f31cc2bc67 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.424774] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 906.432394] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 906.432394] env[62000]: value = "task-882537" [ 906.432394] env[62000]: _type = "Task" [ 906.432394] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.441302] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882537, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.682455] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a98ce5db-a9be-4524-96a6-9bed05348773 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "edc7c99e-d53c-4e0b-991d-c4b693544cdd" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.052s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.700957] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.701452] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Instance network_info: |[{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 906.702814] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.702947] env[62000]: DEBUG nova.network.neutron [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 906.704755] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:9a:5f:74', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a15de394-0367-4921-a5c1-6ac8615e3283', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 906.713717] env[62000]: DEBUG oslo.service.loopingcall [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.716583] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 906.717535] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b492660-190c-4171-b1b8-f7909d6a15e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.738386] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 906.738386] env[62000]: value = "task-882538" [ 906.738386] env[62000]: _type = "Task" [ 906.738386] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.746755] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882538, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.949025] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882537, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.093614} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.949325] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.949549] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 906.949835] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 906.957215] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.965602] env[62000]: DEBUG nova.network.neutron [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updated VIF entry in instance network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 906.965975] env[62000]: DEBUG nova.network.neutron [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.248850] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882538, 'name': CreateVM_Task, 'duration_secs': 0.371647} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.251769] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 907.252785] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.252946] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.255076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 907.255076] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-afccea25-9cbb-4d06-99f0-88aea8380d59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.259487] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 907.259487] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a90306-93df-38e9-8ff8-2847100fde55" [ 907.259487] env[62000]: _type = "Task" [ 907.259487] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.271173] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a90306-93df-38e9-8ff8-2847100fde55, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.384712] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16b4732-52bf-44bc-98e8-04d2bb91e2b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.393258] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865ace57-80d5-470d-b3f2-bc70e86642ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.425800] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5186a9d9-4192-4f3a-8440-c517db0adcbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.432541] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70e7bd1-ab55-4e7d-83ec-ff6e0fbd81a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.446675] env[62000]: DEBUG nova.compute.provider_tree [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.469348] env[62000]: DEBUG oslo_concurrency.lockutils [req-c2973013-d63e-4e68-abf7-e8bd39b23b2f req-9292ff45-b97a-401a-b120-9fb8cf10c042 service nova] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.770811] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a90306-93df-38e9-8ff8-2847100fde55, 'name': SearchDatastore_Task, 'duration_secs': 0.013123} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.771170] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.771485] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 907.771772] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 907.772013] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 907.772226] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 907.772523] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e916270-488c-4490-8f1f-419db8c36a6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.780360] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 907.780554] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 907.781580] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d22bd20e-9b74-4b6d-bac4-71f6cf66d62c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.786419] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 907.786419] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52677082-90e5-f3c7-b6ca-46a7cfad43b0" [ 907.786419] env[62000]: _type = "Task" [ 907.786419] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.793970] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52677082-90e5-f3c7-b6ca-46a7cfad43b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.950114] env[62000]: DEBUG nova.scheduler.client.report [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.979293] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 907.979563] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 907.979725] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 907.979955] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 907.980194] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 907.980360] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 907.980634] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 907.980863] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 907.980977] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 907.981402] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 907.981402] env[62000]: DEBUG nova.virt.hardware [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 907.982728] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-975a2b6e-e1f2-4a6c-9a48-13cba131555e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.991575] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-106cd8c9-6f29-45a9-925a-689acd8796d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.005600] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance VIF info [] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 908.011245] env[62000]: DEBUG oslo.service.loopingcall [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.011490] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 908.011698] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8eefe359-de5d-41da-aa11-7c429c279da6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.029646] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 908.029646] env[62000]: value = "task-882539" [ 908.029646] env[62000]: _type = "Task" [ 908.029646] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.040181] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882539, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.207574] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.207827] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.300942] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52677082-90e5-f3c7-b6ca-46a7cfad43b0, 'name': SearchDatastore_Task, 'duration_secs': 0.007623} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.301891] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2bd508c-028f-4a59-bf24-a03af45a488b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.308186] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 908.308186] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8f3cf-2c12-fedf-f4c4-ec8fed3d5b9f" [ 908.308186] env[62000]: _type = "Task" [ 908.308186] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.321046] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8f3cf-2c12-fedf-f4c4-ec8fed3d5b9f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.455740] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.304s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.456174] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.459338] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.494s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.460804] env[62000]: INFO nova.compute.claims [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.540207] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882539, 'name': CreateVM_Task, 'duration_secs': 0.428396} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.540410] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 908.540850] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.541022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.541352] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 908.541629] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-11652216-3cbd-4ca4-a016-6bad2f1a0410 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.547169] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 908.547169] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525f4bef-94d0-7dd7-0ec0-bdbd79722347" [ 908.547169] env[62000]: _type = "Task" [ 908.547169] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.555364] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525f4bef-94d0-7dd7-0ec0-bdbd79722347, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.710159] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 908.821167] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f8f3cf-2c12-fedf-f4c4-ec8fed3d5b9f, 'name': SearchDatastore_Task, 'duration_secs': 0.011501} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.821518] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.821767] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 94605ab8-8167-430e-b1cd-c8f51e50d8b9/94605ab8-8167-430e-b1cd-c8f51e50d8b9.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 908.822157] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-25d37d05-d2ca-43e6-b872-f775473a39f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.828914] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 908.828914] env[62000]: value = "task-882540" [ 908.828914] env[62000]: _type = "Task" [ 908.828914] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.836276] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.966035] env[62000]: DEBUG nova.compute.utils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 908.969846] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 908.970055] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 909.029149] env[62000]: DEBUG nova.policy [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d0b65722357449ab64736e5e1eb713f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ce3270692934d2c9c1330a45c0e059e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.060859] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525f4bef-94d0-7dd7-0ec0-bdbd79722347, 'name': SearchDatastore_Task, 'duration_secs': 0.010458} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.061180] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.061810] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 909.061810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.062059] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.062059] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 909.063049] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c57c525d-55ba-409a-8c1d-a8b441836f9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.075784] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 909.075993] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 909.076756] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1a5eef94-cdbf-4ea5-b458-c365351f7d8d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.082321] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 909.082321] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dfea8e-19eb-1f54-6c83-d69c0099f95e" [ 909.082321] env[62000]: _type = "Task" [ 909.082321] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.092179] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dfea8e-19eb-1f54-6c83-d69c0099f95e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.234850] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.341443] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.486335} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.341756] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 94605ab8-8167-430e-b1cd-c8f51e50d8b9/94605ab8-8167-430e-b1cd-c8f51e50d8b9.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 909.342019] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 909.342352] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ba766570-ae05-4a39-82c3-ddde7a89b3b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.349032] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 909.349032] env[62000]: value = "task-882541" [ 909.349032] env[62000]: _type = "Task" [ 909.349032] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.357282] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.471344] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.495063] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Successfully created port: 73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.603302] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dfea8e-19eb-1f54-6c83-d69c0099f95e, 'name': SearchDatastore_Task, 'duration_secs': 0.054224} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.608054] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-116d743d-60e9-49e3-91cc-fda04da60bff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.614945] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 909.614945] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52096953-7aa4-62aa-9770-03a8bb312f12" [ 909.614945] env[62000]: _type = "Task" [ 909.614945] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.629719] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52096953-7aa4-62aa-9770-03a8bb312f12, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.728437] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0642e7c6-9252-4dee-a629-93a309c5fb3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.736125] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f456bd44-b56c-4ed2-8808-f6133c3eee7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.766806] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4815438-c856-4ed4-aa23-fe9cdb304af7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.774831] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a698c35-4a4d-4614-8d17-9e2bda16209c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.788836] env[62000]: DEBUG nova.compute.provider_tree [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.858758] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.336036} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.859137] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 909.859848] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-252ded71-9b4b-4326-b8d4-f981cdd021cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.881960] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] 94605ab8-8167-430e-b1cd-c8f51e50d8b9/94605ab8-8167-430e-b1cd-c8f51e50d8b9.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 909.882341] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73e90114-3d8f-43e4-9817-87615c3f29b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.903097] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 909.903097] env[62000]: value = "task-882542" [ 909.903097] env[62000]: _type = "Task" [ 909.903097] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.912824] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882542, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.126794] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52096953-7aa4-62aa-9770-03a8bb312f12, 'name': SearchDatastore_Task, 'duration_secs': 0.013361} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.127171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.127887] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 910.127887] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-779ab35f-693d-40a7-b1c3-c3421796fe24 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.134499] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 910.134499] env[62000]: value = "task-882543" [ 910.134499] env[62000]: _type = "Task" [ 910.134499] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.142892] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882543, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.292834] env[62000]: DEBUG nova.scheduler.client.report [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.417304] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882542, 'name': ReconfigVM_Task, 'duration_secs': 0.283999} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.417899] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfigured VM instance instance-00000052 to attach disk [datastore2] 94605ab8-8167-430e-b1cd-c8f51e50d8b9/94605ab8-8167-430e-b1cd-c8f51e50d8b9.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 910.419235] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f0b4c43-0f57-46f9-94b3-b0ee57ed12eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.432090] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 910.432090] env[62000]: value = "task-882544" [ 910.432090] env[62000]: _type = "Task" [ 910.432090] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.444181] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882544, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.485559] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.512189] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.512754] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.513221] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.513656] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.513970] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.514764] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.514969] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.515311] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.515600] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.516232] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.516232] env[62000]: DEBUG nova.virt.hardware [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.517115] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b5fb1a1-2d2b-4cc4-b678-046bf9780ef4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.529910] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cf4354c-0ab0-4eb1-9276-0ff0a9e76126 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.645748] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882543, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.798079] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.798652] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.801454] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.320s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.803546] env[62000]: INFO nova.compute.claims [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 910.942514] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882544, 'name': Rename_Task, 'duration_secs': 0.292971} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.942823] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 910.944009] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f8814f1d-2183-496e-b990-d765ae7b28a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.949823] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 910.949823] env[62000]: value = "task-882545" [ 910.949823] env[62000]: _type = "Task" [ 910.949823] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.957154] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882545, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.978854] env[62000]: DEBUG nova.compute.manager [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Received event network-vif-plugged-73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 910.978953] env[62000]: DEBUG oslo_concurrency.lockutils [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] Acquiring lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.979144] env[62000]: DEBUG oslo_concurrency.lockutils [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.979318] env[62000]: DEBUG oslo_concurrency.lockutils [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.979487] env[62000]: DEBUG nova.compute.manager [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] No waiting events found dispatching network-vif-plugged-73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.979684] env[62000]: WARNING nova.compute.manager [req-4b9a2941-e0d1-421e-9648-3bea9aa8ca07 req-8faae788-ca5d-4c95-ae19-64adb1da3515 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Received unexpected event network-vif-plugged-73dc270a-70b4-4343-a230-1102c85ce9e6 for instance with vm_state building and task_state spawning. [ 911.089049] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Successfully updated port: 73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.145595] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882543, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.563527} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.145876] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 911.146115] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 911.146378] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-56d283f6-85c8-445c-9f99-a97d0c5352ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.152868] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 911.152868] env[62000]: value = "task-882546" [ 911.152868] env[62000]: _type = "Task" [ 911.152868] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.162686] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882546, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.308105] env[62000]: DEBUG nova.compute.utils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.311340] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.315109] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 911.360423] env[62000]: DEBUG nova.policy [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4770851cc3b4ad8aaa6866a18ee1359', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41e10f7d4f964f2795dc629721802880', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.460452] env[62000]: DEBUG oslo_vmware.api [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882545, 'name': PowerOnVM_Task, 'duration_secs': 0.470393} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.460731] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 911.460933] env[62000]: INFO nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Took 6.95 seconds to spawn the instance on the hypervisor. [ 911.461139] env[62000]: DEBUG nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.461912] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5954456b-081b-4ec9-b14f-53bf86cd9585 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.592181] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.592455] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.592912] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 911.663226] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882546, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.07877} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.663719] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 911.664790] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3924fa80-28aa-4b79-8c56-f3b435a985f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.685557] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 911.686056] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-906ca85d-6829-4324-b0fc-28cb36769b93 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.708733] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 911.708733] env[62000]: value = "task-882547" [ 911.708733] env[62000]: _type = "Task" [ 911.708733] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.716616] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882547, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.736732] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Successfully created port: 65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.820855] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.979045] env[62000]: INFO nova.compute.manager [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Took 20.40 seconds to build instance. [ 912.097573] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4837ce8-986f-4f60-922a-c964863c0038 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.105941] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03279bc-bee6-46f0-b605-dc7942e53575 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.139735] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 912.142400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1759b03c-eee8-48c7-aadb-9ee4c3a162e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.150162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e588fce-443a-4666-ba80-c1a0a895407b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.163979] env[62000]: DEBUG nova.compute.provider_tree [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.220076] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882547, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.355956] env[62000]: DEBUG nova.network.neutron [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updating instance_info_cache with network_info: [{"id": "73dc270a-70b4-4343-a230-1102c85ce9e6", "address": "fa:16:3e:81:15:f6", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73dc270a-70", "ovs_interfaceid": "73dc270a-70b4-4343-a230-1102c85ce9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.479973] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eef9bb8c-205a-47e8-8aae-a85f4c29fe96 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.927s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.671504] env[62000]: DEBUG nova.scheduler.client.report [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.719748] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882547, 'name': ReconfigVM_Task, 'duration_secs': 0.512992} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.719971] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Reconfigured VM instance instance-00000051 to attach disk [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283/2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 912.720648] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bd970bab-9479-401a-bebb-49ed87bd4afc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.727753] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 912.727753] env[62000]: value = "task-882548" [ 912.727753] env[62000]: _type = "Task" [ 912.727753] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.736403] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882548, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.831282] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.861502] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.861798] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.861963] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.862410] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.862553] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.862709] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.862922] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.863130] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.863311] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.863486] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.863649] env[62000]: DEBUG nova.virt.hardware [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.864139] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.864430] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Instance network_info: |[{"id": "73dc270a-70b4-4343-a230-1102c85ce9e6", "address": "fa:16:3e:81:15:f6", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73dc270a-70", "ovs_interfaceid": "73dc270a-70b4-4343-a230-1102c85ce9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 912.865250] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b95c2c2-313c-4248-9756-07d8c6cf885f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.868119] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:15:f6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '73dc270a-70b4-4343-a230-1102c85ce9e6', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 912.875546] env[62000]: DEBUG oslo.service.loopingcall [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 912.875775] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 912.876512] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31ca884b-b810-4108-a93e-b4144b382f4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.894710] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90794c87-f839-488e-808c-0e6d2557a91a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.899719] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 912.899719] env[62000]: value = "task-882549" [ 912.899719] env[62000]: _type = "Task" [ 912.899719] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.915652] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882549, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.008055] env[62000]: DEBUG nova.compute.manager [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Received event network-changed-73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 913.008315] env[62000]: DEBUG nova.compute.manager [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Refreshing instance network info cache due to event network-changed-73dc270a-70b4-4343-a230-1102c85ce9e6. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 913.008491] env[62000]: DEBUG oslo_concurrency.lockutils [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] Acquiring lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.008635] env[62000]: DEBUG oslo_concurrency.lockutils [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] Acquired lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.008800] env[62000]: DEBUG nova.network.neutron [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Refreshing network info cache for port 73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 913.175274] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.175969] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.180291] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.630s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.184019] env[62000]: DEBUG nova.objects.instance [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'resources' on Instance uuid 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.238606] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882548, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.410327] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882549, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.683419] env[62000]: DEBUG nova.objects.instance [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'numa_topology' on Instance uuid 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.685666] env[62000]: DEBUG nova.compute.utils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.686864] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.687045] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 913.719042] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Successfully updated port: 65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.744517] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882548, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.824710] env[62000]: DEBUG nova.policy [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 913.911913] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882549, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.000113] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 914.000565] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78569852-3a4d-4220-ae9d-254cf1b271a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.009200] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 914.009200] env[62000]: ERROR oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk due to incomplete transfer. [ 914.009200] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-c3311b26-6086-477b-805a-1bba81ff96db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.016031] env[62000]: DEBUG oslo_vmware.rw_handles [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52cbec31-5515-9602-6a5c-06990771ae8c/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 914.016368] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Uploaded image d19f59f0-69e8-45e4-9559-cc7320fda017 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 914.018918] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 914.019330] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e6d6fc31-1c17-4cd5-a5ac-2d938ad3d7d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.028032] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 914.028032] env[62000]: value = "task-882550" [ 914.028032] env[62000]: _type = "Task" [ 914.028032] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.035643] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882550, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.188483] env[62000]: DEBUG nova.objects.base [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Object Instance<8a11689f-fc00-43f8-9215-8d81daa84400> lazy-loaded attributes: resources,numa_topology {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 914.192813] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.222835] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.223082] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.223269] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 914.229161] env[62000]: DEBUG nova.network.neutron [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updated VIF entry in instance network info cache for port 73dc270a-70b4-4343-a230-1102c85ce9e6. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 914.229558] env[62000]: DEBUG nova.network.neutron [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updating instance_info_cache with network_info: [{"id": "73dc270a-70b4-4343-a230-1102c85ce9e6", "address": "fa:16:3e:81:15:f6", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73dc270a-70", "ovs_interfaceid": "73dc270a-70b4-4343-a230-1102c85ce9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.243563] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882548, 'name': Rename_Task, 'duration_secs': 1.169302} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.243860] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 914.244116] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-09525539-5ad4-479b-afd8-be04b42d107d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.254370] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 914.254370] env[62000]: value = "task-882551" [ 914.254370] env[62000]: _type = "Task" [ 914.254370] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.265621] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882551, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.412744] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882549, 'name': CreateVM_Task, 'duration_secs': 1.327611} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.413064] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 914.413884] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.414203] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.414625] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 914.414991] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-77e9a3cf-e448-4706-bd45-bded239ccffc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.425017] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 914.425017] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5290b264-13a1-ccfa-97b1-b37caca31344" [ 914.425017] env[62000]: _type = "Task" [ 914.425017] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.432561] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5290b264-13a1-ccfa-97b1-b37caca31344, 'name': SearchDatastore_Task, 'duration_secs': 0.00865} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.435165] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.435644] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.435974] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.436250] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.436541] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.437053] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54149255-e53b-4d31-a233-ef45a9f09d35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.448025] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.448025] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 914.448025] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b766365a-995b-4f70-9ee0-b0ebcd8ec7c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.459016] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 914.459016] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52371580-925c-28c0-b82d-96f62645c212" [ 914.459016] env[62000]: _type = "Task" [ 914.459016] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.465413] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52371580-925c-28c0-b82d-96f62645c212, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.487160] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6556680-af7d-40e6-bf65-b7869d8f1d60 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.492675] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f0eb2f-91d4-470a-af2b-68ee94280b6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.533280] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd04a942-bc91-4799-856f-37acf5fe61f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.548301] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261909d5-cdfd-480d-8163-973334ce81a1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.551705] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882550, 'name': Destroy_Task, 'duration_secs': 0.328206} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.552399] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Destroyed the VM [ 914.552399] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 914.552912] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6468db5d-405b-4caa-81b2-4040f1110ed0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.562988] env[62000]: DEBUG nova.compute.provider_tree [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.572331] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 914.572331] env[62000]: value = "task-882552" [ 914.572331] env[62000]: _type = "Task" [ 914.572331] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.580971] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882552, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.581748] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Successfully created port: 060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.741114] env[62000]: DEBUG oslo_concurrency.lockutils [req-21ee1a6c-cd66-4a59-ab4b-4236df7bb3aa req-907746f9-cd01-474a-823b-235b50dc0b8f service nova] Releasing lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.765599] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882551, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.769437] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 914.969045] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52371580-925c-28c0-b82d-96f62645c212, 'name': SearchDatastore_Task, 'duration_secs': 0.010604} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.969851] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f48e8f2-0731-47c5-9041-1077fb44cb8f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.975180] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 914.975180] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b21acd-afd8-23c5-c3d4-56da29685698" [ 914.975180] env[62000]: _type = "Task" [ 914.975180] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.983328] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b21acd-afd8-23c5-c3d4-56da29685698, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.055574] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 915.055574] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 915.055784] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.055819] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.055996] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 915.059848] env[62000]: DEBUG nova.network.neutron [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.067031] env[62000]: DEBUG nova.scheduler.client.report [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.084452] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882552, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.178269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "54eb64ad-bc7b-4495-9e93-451059139c4c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.178269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.212242] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.236564] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.236809] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.237133] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.237238] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.237325] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.237457] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.237664] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.237880] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.238138] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.238251] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.238425] env[62000]: DEBUG nova.virt.hardware [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.239358] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04f23f8c-4c8f-4bce-8f53-155d749bfa39 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.247650] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d83d92d-26f9-42ff-a6ad-8d7b20d79a13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.268296] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882551, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.485914] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b21acd-afd8-23c5-c3d4-56da29685698, 'name': SearchDatastore_Task, 'duration_secs': 0.009641} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.486366] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.486645] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a0e2344f-32ec-4fb1-ba7c-99e18f422923/a0e2344f-32ec-4fb1-ba7c-99e18f422923.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 915.486914] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-571116d4-f680-4109-b0ae-8175718e8c13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.492959] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 915.492959] env[62000]: value = "task-882553" [ 915.492959] env[62000]: _type = "Task" [ 915.492959] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.500405] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882553, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.562497] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.562876] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Instance network_info: |[{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.563366] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:41:7c:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '65be2bd7-e84a-4cc8-b937-6140dc13b2e4', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.572583] env[62000]: DEBUG oslo.service.loopingcall [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.572862] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 915.573719] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.394s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.576671] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8b189fd3-cc03-4645-b5fa-185134fdc866 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.593879] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.285s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.594174] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.597375] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.640s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.599312] env[62000]: INFO nova.compute.claims [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 915.623818] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882552, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.625772] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.625772] env[62000]: value = "task-882554" [ 915.625772] env[62000]: _type = "Task" [ 915.625772] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.635359] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882554, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.651625] env[62000]: INFO nova.scheduler.client.report [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocations for instance 0a3be3e8-b079-4006-8a46-9b9dd02baa5b [ 915.680871] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 915.775539] env[62000]: DEBUG oslo_vmware.api [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882551, 'name': PowerOnVM_Task, 'duration_secs': 1.229267} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.775949] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 915.776276] env[62000]: DEBUG nova.compute.manager [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 915.777306] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de8a49cb-d476-421b-944c-a976d3f45d14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.928343] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 915.928897] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.003443] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882553, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.085678] env[62000]: DEBUG oslo_vmware.api [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882552, 'name': RemoveSnapshot_Task, 'duration_secs': 1.44244} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.086405] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 916.086715] env[62000]: INFO nova.compute.manager [None req-82592792-7bcd-4f00-ab8f-35ee471c6a80 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Took 17.63 seconds to snapshot the instance on the hypervisor. [ 916.125523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fe2e8188-abfe-444c-8876-5506f04e645a tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 34.060s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.126444] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 11.098s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.126649] env[62000]: INFO nova.compute.manager [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Unshelving [ 916.147350] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882554, 'name': CreateVM_Task, 'duration_secs': 0.377602} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.147350] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 916.147350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.147350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.147350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.147350] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3d28f434-47b5-48ca-9c97-3f4d6de8fc7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.154735] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 916.154735] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f71096-9c5f-4a68-cad2-7eaff610c01e" [ 916.154735] env[62000]: _type = "Task" [ 916.154735] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.159762] env[62000]: DEBUG oslo_concurrency.lockutils [None req-b4798609-78d5-45d8-98d3-5000cfe00f04 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0a3be3e8-b079-4006-8a46-9b9dd02baa5b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.779s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.171897] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f71096-9c5f-4a68-cad2-7eaff610c01e, 'name': SearchDatastore_Task, 'duration_secs': 0.008612} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.172923] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.173262] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 916.173566] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.173778] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.174033] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 916.174562] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-adaaaa43-65b5-49ab-b810-6487de01da72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.183239] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 916.183513] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 916.188573] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3665b649-230d-4a0d-aa0c-849b2dcb479d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.194848] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 916.194848] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238e039-4c51-2173-2db1-2e93e937ac06" [ 916.194848] env[62000]: _type = "Task" [ 916.194848] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.203471] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238e039-4c51-2173-2db1-2e93e937ac06, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.207684] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.300027] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.302417] env[62000]: DEBUG nova.compute.manager [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Received event network-vif-plugged-060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.302706] env[62000]: DEBUG oslo_concurrency.lockutils [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] Acquiring lock "bf606000-346b-48db-972a-000b54a8ec5d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.303078] env[62000]: DEBUG oslo_concurrency.lockutils [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] Lock "bf606000-346b-48db-972a-000b54a8ec5d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.303341] env[62000]: DEBUG oslo_concurrency.lockutils [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] Lock "bf606000-346b-48db-972a-000b54a8ec5d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.303611] env[62000]: DEBUG nova.compute.manager [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] No waiting events found dispatching network-vif-plugged-060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 916.303845] env[62000]: WARNING nova.compute.manager [req-01c22c04-5e8d-4e50-a5c9-748922cf7202 req-b7a3a8b1-9924-42f8-a690-0b7e48c1867a service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Received unexpected event network-vif-plugged-060c8648-df77-440f-81c1-8d8d84a20a5d for instance with vm_state building and task_state spawning. [ 916.425596] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Successfully updated port: 060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.431468] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.431736] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 916.431918] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing instance network info cache due to event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 916.432238] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.432388] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.432553] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 916.503727] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882553, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.523897} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.503966] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a0e2344f-32ec-4fb1-ba7c-99e18f422923/a0e2344f-32ec-4fb1-ba7c-99e18f422923.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 916.504212] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 916.504464] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fd5843fc-7ab6-45c4-8ac5-3db36184514a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.510896] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 916.510896] env[62000]: value = "task-882555" [ 916.510896] env[62000]: _type = "Task" [ 916.510896] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.519811] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.707634] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5238e039-4c51-2173-2db1-2e93e937ac06, 'name': SearchDatastore_Task, 'duration_secs': 0.008613} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.708599] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-41f100cb-ac45-40d8-9133-9739cf6e3cba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.713471] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 916.713471] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee1586-aefd-b352-3cea-70e9ad062119" [ 916.713471] env[62000]: _type = "Task" [ 916.713471] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.723140] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee1586-aefd-b352-3cea-70e9ad062119, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.821867] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7544c8af-d2bf-4912-8bb8-1646aad051b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.829392] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-985a2edb-edcc-4865-8e87-d10878a0eeac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.858252] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca16ac9f-0eb4-412b-9aa4-da7978d2b054 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.865452] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94eeafb0-f56f-4fe3-9cab-e3500da6a83d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.879967] env[62000]: DEBUG nova.compute.provider_tree [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.897719] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.897964] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.898198] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 916.898387] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 916.898554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.900556] env[62000]: INFO nova.compute.manager [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Terminating instance [ 916.902262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "refresh_cache-2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.902428] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "refresh_cache-2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.902596] env[62000]: DEBUG nova.network.neutron [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 916.928519] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.928669] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.928808] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 917.025307] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061132} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.025620] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.026435] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75526009-af34-4b49-a5af-d5e6c75f8838 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.051763] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] a0e2344f-32ec-4fb1-ba7c-99e18f422923/a0e2344f-32ec-4fb1-ba7c-99e18f422923.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 917.054389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b2f18aed-3c02-4221-8936-222afaf7f367 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.074768] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 917.074768] env[62000]: value = "task-882556" [ 917.074768] env[62000]: _type = "Task" [ 917.074768] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.082546] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.086320] env[62000]: DEBUG nova.compute.manager [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.086320] env[62000]: DEBUG nova.compute.manager [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing instance network info cache due to event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.086588] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.149258] env[62000]: DEBUG nova.compute.utils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 917.230030] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee1586-aefd-b352-3cea-70e9ad062119, 'name': SearchDatastore_Task, 'duration_secs': 0.00985} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.230030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.232097] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 917.232097] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9735aa93-8a63-4995-9b12-09701e2fb5fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.239746] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 917.239746] env[62000]: value = "task-882557" [ 917.239746] env[62000]: _type = "Task" [ 917.239746] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.250584] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882557, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.383484] env[62000]: DEBUG nova.scheduler.client.report [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 917.403412] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updated VIF entry in instance network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 917.403813] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.420488] env[62000]: DEBUG nova.network.neutron [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 917.476393] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 917.517126] env[62000]: DEBUG nova.network.neutron [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.586755] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882556, 'name': ReconfigVM_Task, 'duration_secs': 0.262447} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.587059] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Reconfigured VM instance instance-00000053 to attach disk [datastore2] a0e2344f-32ec-4fb1-ba7c-99e18f422923/a0e2344f-32ec-4fb1-ba7c-99e18f422923.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 917.587735] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6f83e29b-e68f-4f9b-b10f-82c33316e92e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.596293] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 917.596293] env[62000]: value = "task-882558" [ 917.596293] env[62000]: _type = "Task" [ 917.596293] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.609249] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882558, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.649612] env[62000]: DEBUG nova.network.neutron [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Updating instance_info_cache with network_info: [{"id": "060c8648-df77-440f-81c1-8d8d84a20a5d", "address": "fa:16:3e:0d:7b:d0", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap060c8648-df", "ovs_interfaceid": "060c8648-df77-440f-81c1-8d8d84a20a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.653102] env[62000]: INFO nova.virt.block_device [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Booting with volume ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5 at /dev/sdb [ 917.692296] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b205b0e4-290e-43e3-884f-333a5342327f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.702059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ad83e3-ac55-4f0d-ba65-e5917773adaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.733070] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-722df479-693a-47da-ae71-54b5ec04ebc2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.740614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acd72e1e-d9bc-4d9b-afcb-319214f792d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.759163] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882557, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.759434] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 917.759650] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.759920] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-008213e4-3714-43fb-bc2d-998794fa5d35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.765776] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 917.765776] env[62000]: value = "task-882559" [ 917.765776] env[62000]: _type = "Task" [ 917.765776] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.778858] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ea0445f-c60f-4b32-a882-46075bb34250 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.786133] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882559, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.788648] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5cc94e1-14eb-4d3d-93c2-3161408ab178 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.803190] env[62000]: DEBUG nova.virt.block_device [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating existing volume attachment record: 450fe1c5-d3c7-40cf-b740-8b954b02dec8 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 917.888771] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.291s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.889383] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 917.892303] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.658s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.893752] env[62000]: INFO nova.compute.claims [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 917.908488] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.908780] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Received event network-vif-plugged-65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.909009] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.909254] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.909436] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.909616] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] No waiting events found dispatching network-vif-plugged-65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.909792] env[62000]: WARNING nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Received unexpected event network-vif-plugged-65be2bd7-e84a-4cc8-b937-6140dc13b2e4 for instance with vm_state building and task_state spawning. [ 917.910042] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Received event network-changed-65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 917.910258] env[62000]: DEBUG nova.compute.manager [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Refreshing instance network info cache due to event network-changed-65be2bd7-e84a-4cc8-b937-6140dc13b2e4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 917.910479] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquiring lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.910630] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Acquired lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.910824] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Refreshing network info cache for port 65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 917.911909] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.912158] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.020682] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "refresh_cache-2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.021050] env[62000]: DEBUG nova.compute.manager [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 918.021342] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 918.022418] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d741a2-83c6-4e5a-a7ab-c95119ae4335 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.030521] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 918.030521] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ef4554d1-fb6f-4ee2-bfeb-73fd55eaedcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.036804] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 918.036804] env[62000]: value = "task-882561" [ 918.036804] env[62000]: _type = "Task" [ 918.036804] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.044730] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882561, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.105610] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882558, 'name': Rename_Task, 'duration_secs': 0.206196} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.105800] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 918.106053] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1ec75473-0ac5-43c8-9849-f8b82be2ef28 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.111937] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 918.111937] env[62000]: value = "task-882562" [ 918.111937] env[62000]: _type = "Task" [ 918.111937] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.119586] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882562, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.152805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.153151] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Instance network_info: |[{"id": "060c8648-df77-440f-81c1-8d8d84a20a5d", "address": "fa:16:3e:0d:7b:d0", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap060c8648-df", "ovs_interfaceid": "060c8648-df77-440f-81c1-8d8d84a20a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.153625] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:7b:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '060c8648-df77-440f-81c1-8d8d84a20a5d', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.161874] env[62000]: DEBUG oslo.service.loopingcall [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.162169] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 918.162411] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5e9b4ae2-3916-4fd9-9f85-3ca08db7c61e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.181324] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.181324] env[62000]: value = "task-882565" [ 918.181324] env[62000]: _type = "Task" [ 918.181324] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.188863] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882565, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.284786] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882559, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064782} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.285104] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 918.285891] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b97bdea3-d60f-4f8f-a53b-fab2ff1f7a5f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.307540] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.307882] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29441119-cbdc-40e7-82be-3c9b1236dcbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.327502] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 918.327502] env[62000]: value = "task-882566" [ 918.327502] env[62000]: _type = "Task" [ 918.327502] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.329909] env[62000]: DEBUG nova.compute.manager [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Received event network-changed-060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 918.329976] env[62000]: DEBUG nova.compute.manager [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Refreshing instance network info cache due to event network-changed-060c8648-df77-440f-81c1-8d8d84a20a5d. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 918.330465] env[62000]: DEBUG oslo_concurrency.lockutils [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] Acquiring lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.330624] env[62000]: DEBUG oslo_concurrency.lockutils [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] Acquired lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.330798] env[62000]: DEBUG nova.network.neutron [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Refreshing network info cache for port 060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 918.340755] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882566, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.398318] env[62000]: DEBUG nova.compute.utils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 918.402058] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 918.402058] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 918.508874] env[62000]: DEBUG nova.policy [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3018dc2fa8d413aa464bf4e0c32341e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d131f95ad3949d89cd6f36f6648d3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 918.548725] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882561, 'name': PowerOffVM_Task, 'duration_secs': 0.270097} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.549069] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 918.549255] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 918.549527] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62cc0a14-58c1-4e83-a09e-f92b327f87fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.581394] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 918.581576] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 918.581815] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleting the datastore file [datastore2] 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.582172] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-20d9b067-2d2c-47e1-9438-4c5d9e68799d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.589138] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 918.589138] env[62000]: value = "task-882568" [ 918.589138] env[62000]: _type = "Task" [ 918.589138] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.603893] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882568, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.621606] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882562, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.644654] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updated VIF entry in instance network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.645142] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.691428] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882565, 'name': CreateVM_Task, 'duration_secs': 0.361295} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.691683] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 918.692337] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.692518] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.692838] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 918.693422] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16fec6c8-2279-4a17-9f56-26f87c9a1138 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.698381] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 918.698381] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525617b6-5f24-9c03-f671-383e7b7139db" [ 918.698381] env[62000]: _type = "Task" [ 918.698381] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.705841] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525617b6-5f24-9c03-f671-383e7b7139db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.840456] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882566, 'name': ReconfigVM_Task, 'duration_secs': 0.332374} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.840711] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfigured VM instance instance-00000054 to attach disk [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.841636] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2a1999f9-fb06-4b78-99ed-11ee77a089d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.845471] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updated VIF entry in instance network info cache for port 65be2bd7-e84a-4cc8-b937-6140dc13b2e4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 918.845808] env[62000]: DEBUG nova.network.neutron [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.848079] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 918.848079] env[62000]: value = "task-882569" [ 918.848079] env[62000]: _type = "Task" [ 918.848079] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.856345] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882569, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.903203] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 918.979744] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Successfully created port: 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 919.048113] env[62000]: DEBUG nova.network.neutron [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Updated VIF entry in instance network info cache for port 060c8648-df77-440f-81c1-8d8d84a20a5d. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 919.048497] env[62000]: DEBUG nova.network.neutron [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Updating instance_info_cache with network_info: [{"id": "060c8648-df77-440f-81c1-8d8d84a20a5d", "address": "fa:16:3e:0d:7b:d0", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap060c8648-df", "ovs_interfaceid": "060c8648-df77-440f-81c1-8d8d84a20a5d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.104702] env[62000]: DEBUG oslo_vmware.api [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882568, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.282972} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.105061] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 919.105326] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 919.105540] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 919.105770] env[62000]: INFO nova.compute.manager [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Took 1.08 seconds to destroy the instance on the hypervisor. [ 919.106162] env[62000]: DEBUG oslo.service.loopingcall [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 919.106439] env[62000]: DEBUG nova.compute.manager [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 919.106542] env[62000]: DEBUG nova.network.neutron [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 919.122387] env[62000]: DEBUG oslo_vmware.api [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882562, 'name': PowerOnVM_Task, 'duration_secs': 0.564818} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.125334] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.125334] env[62000]: INFO nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Took 8.64 seconds to spawn the instance on the hypervisor. [ 919.125427] env[62000]: DEBUG nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.126449] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf3528cd-fa1e-475b-aaf0-c0e4f1f223d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.129455] env[62000]: DEBUG nova.network.neutron [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.152326] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.152326] env[62000]: DEBUG nova.compute.manager [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 919.152326] env[62000]: DEBUG nova.compute.manager [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 919.152326] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.152326] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.152326] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 919.160019] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33a62f48-32d0-4244-8dff-fef463e4dd59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.169444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62a5cb20-754f-4023-bea0-a0e24749b709 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.206892] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c0644b-af6c-4107-b0c3-c7ab5eee1057 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.216228] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525617b6-5f24-9c03-f671-383e7b7139db, 'name': SearchDatastore_Task, 'duration_secs': 0.062364} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.218791] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.219071] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 919.219380] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.219694] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.219919] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 919.220323] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4adf836c-cfc2-4306-9d54-6b305ebae7b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.223680] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35c4b164-2efc-4f6f-9ec6-3c44bc30fa7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.239653] env[62000]: DEBUG nova.compute.provider_tree [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 919.242059] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 919.242304] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 919.243396] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cbf6faed-90a0-40d1-b2ff-220048feeaf0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.249756] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 919.249756] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5273ca72-96a6-b221-0b37-865a79c092eb" [ 919.249756] env[62000]: _type = "Task" [ 919.249756] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.258617] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5273ca72-96a6-b221-0b37-865a79c092eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.349792] env[62000]: DEBUG oslo_concurrency.lockutils [req-834dadd8-512a-4efe-ab21-9b8f7372fdb6 req-b083a42a-f830-4c7e-bbda-a41b97438741 service nova] Releasing lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.359241] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882569, 'name': Rename_Task, 'duration_secs': 0.325804} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.359522] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 919.359779] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fe2f347a-9936-4aac-80d7-efbe95985c4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.366055] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 919.366055] env[62000]: value = "task-882570" [ 919.366055] env[62000]: _type = "Task" [ 919.366055] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.374401] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882570, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.553046] env[62000]: DEBUG oslo_concurrency.lockutils [req-308c311d-8fa7-4dcb-9bcb-03e8e41f9132 req-d00cb6d7-cc1c-44c9-a1ee-2cfc94b053f2 service nova] Releasing lock "refresh_cache-bf606000-346b-48db-972a-000b54a8ec5d" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.632274] env[62000]: DEBUG nova.network.neutron [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.649072] env[62000]: INFO nova.compute.manager [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Took 17.34 seconds to build instance. [ 919.744841] env[62000]: DEBUG nova.scheduler.client.report [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.760197] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5273ca72-96a6-b221-0b37-865a79c092eb, 'name': SearchDatastore_Task, 'duration_secs': 0.015907} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.760970] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ce66aec-999c-4dc7-9eb6-8d450939ba13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.768054] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 919.768054] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521b36d8-1faa-3746-3be9-031261eec4de" [ 919.768054] env[62000]: _type = "Task" [ 919.768054] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.780760] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521b36d8-1faa-3746-3be9-031261eec4de, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.875521] env[62000]: DEBUG oslo_vmware.api [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882570, 'name': PowerOnVM_Task, 'duration_secs': 0.471695} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.879020] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 919.879020] env[62000]: INFO nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Took 7.04 seconds to spawn the instance on the hypervisor. [ 919.879020] env[62000]: DEBUG nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.879020] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bf85cc-9d9e-48c0-9a03-12dd16efa237 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.918531] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 919.945517] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 919.945795] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 919.945982] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 919.946189] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 919.946343] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 919.946493] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 919.946700] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 919.946864] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 919.947041] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 919.947207] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 919.947407] env[62000]: DEBUG nova.virt.hardware [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 919.948278] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e5bafc-03d1-4708-b6a0-c4b34a6bf9ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.958459] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d0d1b8-7e30-4467-a6da-728f21fe830e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.033034] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 920.033459] env[62000]: DEBUG nova.network.neutron [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.135085] env[62000]: INFO nova.compute.manager [-] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Took 1.03 seconds to deallocate network for instance. [ 920.150286] env[62000]: DEBUG oslo_concurrency.lockutils [None req-56ae39c3-0ee7-4d73-bb95-24df509d1e28 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.856s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.249787] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.357s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.250156] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 920.253064] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.045s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.254213] env[62000]: INFO nova.compute.claims [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 920.277880] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521b36d8-1faa-3746-3be9-031261eec4de, 'name': SearchDatastore_Task, 'duration_secs': 0.015588} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.278198] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.278424] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] bf606000-346b-48db-972a-000b54a8ec5d/bf606000-346b-48db-972a-000b54a8ec5d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 920.278688] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-de30dd26-5275-4da4-90b5-3df3c4049b53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.285197] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 920.285197] env[62000]: value = "task-882572" [ 920.285197] env[62000]: _type = "Task" [ 920.285197] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.298474] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882572, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.393839] env[62000]: INFO nova.compute.manager [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Took 17.44 seconds to build instance. [ 920.541129] env[62000]: DEBUG oslo_concurrency.lockutils [req-9f1f6d4e-ce19-4302-ac97-e5f08316b91a req-5f884358-5fa3-495f-b188-7352e0ff2cb0 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.627177] env[62000]: DEBUG nova.compute.manager [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Received event network-vif-plugged-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.627542] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.627789] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.627977] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.628231] env[62000]: DEBUG nova.compute.manager [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] No waiting events found dispatching network-vif-plugged-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 920.628417] env[62000]: WARNING nova.compute.manager [req-7f8078bb-c9ed-4274-ad93-19bb85d67ca2 req-a1c12397-f318-473a-b2ca-a1b6f29365eb service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Received unexpected event network-vif-plugged-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 for instance with vm_state building and task_state spawning. [ 920.642555] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.648942] env[62000]: DEBUG nova.compute.manager [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Received event network-changed-73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 920.649658] env[62000]: DEBUG nova.compute.manager [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Refreshing instance network info cache due to event network-changed-73dc270a-70b4-4343-a230-1102c85ce9e6. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 920.649926] env[62000]: DEBUG oslo_concurrency.lockutils [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] Acquiring lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.650389] env[62000]: DEBUG oslo_concurrency.lockutils [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] Acquired lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.650664] env[62000]: DEBUG nova.network.neutron [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Refreshing network info cache for port 73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 920.740869] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Successfully updated port: 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 920.761631] env[62000]: DEBUG nova.compute.utils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 920.767165] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 920.767359] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 920.805022] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882572, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.823058] env[62000]: DEBUG nova.policy [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 920.895822] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5fa0d2a-32d8-4c03-bc6e-48372f93cf03 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.954s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.215570] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Successfully created port: e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 921.243390] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.243604] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.243776] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 921.278072] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 921.301460] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882572, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636489} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.304693] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] bf606000-346b-48db-972a-000b54a8ec5d/bf606000-346b-48db-972a-000b54a8ec5d.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 921.304986] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 921.306420] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8b2bcc79-3f5f-469b-a314-a1f555eaf2a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.313697] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 921.313697] env[62000]: value = "task-882573" [ 921.313697] env[62000]: _type = "Task" [ 921.313697] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.324710] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.460959] env[62000]: DEBUG nova.network.neutron [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updated VIF entry in instance network info cache for port 73dc270a-70b4-4343-a230-1102c85ce9e6. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 921.461370] env[62000]: DEBUG nova.network.neutron [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updating instance_info_cache with network_info: [{"id": "73dc270a-70b4-4343-a230-1102c85ce9e6", "address": "fa:16:3e:81:15:f6", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.178", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap73dc270a-70", "ovs_interfaceid": "73dc270a-70b4-4343-a230-1102c85ce9e6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.524292] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b27b5df3-c725-444b-bdee-c892e5e7c87a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.532146] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acd570e-7cd8-4aff-80a3-f6cec6523406 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.562539] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126c8d96-281f-4fd2-a963-9cf021bf014f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.570175] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d696222-1b7c-4587-8eea-9b4a9fbf4e4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.585221] env[62000]: DEBUG nova.compute.provider_tree [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 921.796164] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 921.829589] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073427} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.829888] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 921.830710] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cbf4ad7-a282-4e3c-a944-fbe99cdcb726 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.854922] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Reconfiguring VM instance instance-00000055 to attach disk [datastore2] bf606000-346b-48db-972a-000b54a8ec5d/bf606000-346b-48db-972a-000b54a8ec5d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 921.857799] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3056393f-e8bb-40eb-be0d-085a73085cca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.878178] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 921.878178] env[62000]: value = "task-882574" [ 921.878178] env[62000]: _type = "Task" [ 921.878178] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.886272] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882574, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.963625] env[62000]: DEBUG oslo_concurrency.lockutils [req-e494d379-1108-409c-92c9-ae17d56c39c3 req-95a64b73-f0b6-43a6-a716-e60523690efd service nova] Releasing lock "refresh_cache-a0e2344f-32ec-4fb1-ba7c-99e18f422923" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.965341] env[62000]: DEBUG nova.network.neutron [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.088267] env[62000]: DEBUG nova.scheduler.client.report [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.234318] env[62000]: DEBUG nova.compute.manager [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 922.287825] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 922.313538] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 922.313812] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 922.313974] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 922.314208] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 922.314380] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 922.314545] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 922.314759] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 922.314922] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 922.315114] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 922.315320] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 922.315513] env[62000]: DEBUG nova.virt.hardware [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 922.316400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7ffedd-13ad-422e-8f02-3e23baa6fa10 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.324105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f1e7a0e-9e9d-4267-92d1-5bd6f398c8d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.387958] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882574, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.467864] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.468168] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Instance network_info: |[{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 922.468601] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:16:ab:6a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1eb9c5c5-8fe1-4652-986f-3b9abdd36485', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 922.476403] env[62000]: DEBUG oslo.service.loopingcall [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 922.476655] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 922.476921] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a0c29c83-daee-4b90-af1c-5749cc3bd4a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.496489] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 922.496489] env[62000]: value = "task-882575" [ 922.496489] env[62000]: _type = "Task" [ 922.496489] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.503925] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882575, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.593028] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.593622] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 922.596261] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 6.297s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.596489] env[62000]: DEBUG nova.objects.instance [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 922.680151] env[62000]: DEBUG nova.compute.manager [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Received event network-changed-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 922.680151] env[62000]: DEBUG nova.compute.manager [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Refreshing instance network info cache due to event network-changed-1eb9c5c5-8fe1-4652-986f-3b9abdd36485. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 922.680151] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.680151] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.680151] env[62000]: DEBUG nova.network.neutron [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Refreshing network info cache for port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 922.755053] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.791566] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Successfully updated port: e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 922.892310] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882574, 'name': ReconfigVM_Task, 'duration_secs': 0.586194} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.892310] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Reconfigured VM instance instance-00000055 to attach disk [datastore2] bf606000-346b-48db-972a-000b54a8ec5d/bf606000-346b-48db-972a-000b54a8ec5d.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 922.892541] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b03c192e-ee28-4626-b608-1b7c5420f7cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.900231] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 922.900231] env[62000]: value = "task-882576" [ 922.900231] env[62000]: _type = "Task" [ 922.900231] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.908893] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882576, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.007039] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882575, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.101980] env[62000]: DEBUG nova.compute.utils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 923.106669] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 923.106868] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 923.152537] env[62000]: DEBUG nova.policy [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffd8327dd8c4844b776982e99de5e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14aec090f2a4abc8078df1fe4428bbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 923.294805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 923.295126] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 923.295126] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 923.310509] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.310770] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.311828] env[62000]: DEBUG nova.objects.instance [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'flavor' on Instance uuid 72a5bf70-dc6e-4887-abb8-8fbad64bb065 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.413384] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882576, 'name': Rename_Task, 'duration_secs': 0.25318} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.413685] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 923.413938] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6df6a88-f27a-49d1-bf52-75ae1ef6cb2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.418574] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.424728] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Successfully created port: 5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 923.427659] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 923.427659] env[62000]: value = "task-882577" [ 923.427659] env[62000]: _type = "Task" [ 923.427659] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.435653] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882577, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.453771] env[62000]: DEBUG nova.network.neutron [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updated VIF entry in instance network info cache for port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 923.453906] env[62000]: DEBUG nova.network.neutron [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 923.506591] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882575, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.611930] env[62000]: DEBUG oslo_concurrency.lockutils [None req-78732f9f-fe9e-4e01-b052-e7d65a89a222 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.016s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.613622] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 923.616414] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.974s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.616641] env[62000]: DEBUG nova.objects.instance [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lazy-loading 'resources' on Instance uuid 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.836421] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 923.911818] env[62000]: DEBUG nova.objects.instance [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'pci_requests' on Instance uuid 72a5bf70-dc6e-4887-abb8-8fbad64bb065 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.939676] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882577, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.956821] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.957142] env[62000]: DEBUG nova.compute.manager [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received event network-vif-plugged-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 923.957352] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.957563] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.957729] env[62000]: DEBUG oslo_concurrency.lockutils [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.957902] env[62000]: DEBUG nova.compute.manager [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] No waiting events found dispatching network-vif-plugged-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 923.958089] env[62000]: WARNING nova.compute.manager [req-e7f3ce55-6289-46e1-aecc-9ae29ddf2332 req-70bb8435-3523-4ced-a4dd-2ebd8bc9a861 service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received unexpected event network-vif-plugged-e041eaad-7b03-44b1-be56-0c3871c66ecc for instance with vm_state building and task_state spawning. [ 924.007532] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882575, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.101711] env[62000]: DEBUG nova.network.neutron [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updating instance_info_cache with network_info: [{"id": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "address": "fa:16:3e:4c:38:15", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape041eaad-7b", "ovs_interfaceid": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 924.408168] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd633c23-e3b2-41d7-87de-ef330a3bb77d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.417480] env[62000]: DEBUG nova.objects.base [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Object Instance<72a5bf70-dc6e-4887-abb8-8fbad64bb065> lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 924.417710] env[62000]: DEBUG nova.network.neutron [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 924.420930] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da2c98f-f843-449a-9a7b-1812c06f43bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.462366] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13a2a7ca-7e29-4c6b-afbd-d55440a4fa08 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.470909] env[62000]: DEBUG oslo_vmware.api [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882577, 'name': PowerOnVM_Task, 'duration_secs': 0.836088} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.473261] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 924.473497] env[62000]: INFO nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Took 9.26 seconds to spawn the instance on the hypervisor. [ 924.473677] env[62000]: DEBUG nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.474591] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f849136b-8cd9-4c72-b836-adfcc71a269a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.478745] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d4f0ab-477e-457c-b39c-cfa48b976969 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.500221] env[62000]: DEBUG nova.compute.provider_tree [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.510274] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882575, 'name': CreateVM_Task, 'duration_secs': 1.920496} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.510451] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 924.511181] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.511358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.511670] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 924.511934] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d5db32a-b499-4107-8dd9-cd04a7f2f49f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.517314] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 924.517314] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ca16f9-3822-466e-57d8-4f0f4986958b" [ 924.517314] env[62000]: _type = "Task" [ 924.517314] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.529093] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ca16f9-3822-466e-57d8-4f0f4986958b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.546181] env[62000]: DEBUG nova.policy [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 924.604205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 924.604555] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance network_info: |[{"id": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "address": "fa:16:3e:4c:38:15", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape041eaad-7b", "ovs_interfaceid": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 924.605014] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4c:38:15', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e041eaad-7b03-44b1-be56-0c3871c66ecc', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 924.612895] env[62000]: DEBUG oslo.service.loopingcall [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.613980] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 924.614237] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d5f97f26-19d8-4855-be6e-6fc916d6767b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.629654] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 924.638663] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 924.638663] env[62000]: value = "task-882578" [ 924.638663] env[62000]: _type = "Task" [ 924.638663] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.647589] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882578, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.657635] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='5f9f02a5d648f99143c2defa2e3ec244',container_format='bare',created_at=2024-09-16T06:59:16Z,direct_url=,disk_format='vmdk',id=d19f59f0-69e8-45e4-9559-cc7320fda017,min_disk=1,min_ram=0,name='tempest-test-snap-228714914',owner='d14aec090f2a4abc8078df1fe4428bbb',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-09-16T06:59:33Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 924.657902] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 924.658085] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 924.658295] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 924.658442] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 924.658593] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 924.658808] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 924.658973] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 924.659366] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 924.659556] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 924.659738] env[62000]: DEBUG nova.virt.hardware [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 924.660713] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4aca46-6a2b-4951-9247-c48a55a6436f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.671777] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e5854b7-64f6-4802-98ab-694c6f0a0b53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.967000] env[62000]: DEBUG nova.compute.manager [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received event network-changed-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 924.967233] env[62000]: DEBUG nova.compute.manager [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Refreshing instance network info cache due to event network-changed-e041eaad-7b03-44b1-be56-0c3871c66ecc. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 924.967753] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] Acquiring lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 924.967753] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] Acquired lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 924.967753] env[62000]: DEBUG nova.network.neutron [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Refreshing network info cache for port e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 924.982827] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Successfully updated port: 5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 925.010697] env[62000]: DEBUG nova.scheduler.client.report [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.014350] env[62000]: INFO nova.compute.manager [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Took 21.55 seconds to build instance. [ 925.032882] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ca16f9-3822-466e-57d8-4f0f4986958b, 'name': SearchDatastore_Task, 'duration_secs': 0.009747} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.033925] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.034191] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.034258] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.034438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.034731] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.035258] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5bf5b619-418a-4ca3-be1b-98823772162d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.046913] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.047136] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 925.047874] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-04f6b289-3f0b-4605-b846-4cda238272b1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.059857] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 925.059857] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52548f3b-86fc-2ac9-5325-1f01ed3f55c1" [ 925.059857] env[62000]: _type = "Task" [ 925.059857] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.071661] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52548f3b-86fc-2ac9-5325-1f01ed3f55c1, 'name': SearchDatastore_Task, 'duration_secs': 0.009931} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.072340] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2d85d1c-8eb1-4981-be75-ffa1a9329659 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.078318] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 925.078318] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b66892-51f1-eeee-a667-c9bc44955bf0" [ 925.078318] env[62000]: _type = "Task" [ 925.078318] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.088592] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b66892-51f1-eeee-a667-c9bc44955bf0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.149197] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882578, 'name': CreateVM_Task, 'duration_secs': 0.331332} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.149397] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 925.150132] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.150311] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.150657] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 925.150904] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-603221b9-e059-4dc1-9acd-9be93ceb178b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.156297] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 925.156297] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d48143-869a-9f6a-e853-dfb354026c18" [ 925.156297] env[62000]: _type = "Task" [ 925.156297] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.164354] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d48143-869a-9f6a-e853-dfb354026c18, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.487072] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.487433] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.487433] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 925.519883] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c27c20b-ee56-488f-b9cd-eb2eb305c8bd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.062s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.520630] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.904s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.523032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 2.768s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.547292] env[62000]: INFO nova.scheduler.client.report [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleted allocations for instance 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283 [ 925.589648] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b66892-51f1-eeee-a667-c9bc44955bf0, 'name': SearchDatastore_Task, 'duration_secs': 0.010673} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.589736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.590097] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 925.591094] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fe7e62d0-e5cf-451b-807a-08a70d633840 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.600854] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 925.600854] env[62000]: value = "task-882579" [ 925.600854] env[62000]: _type = "Task" [ 925.600854] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.611200] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882579, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.668586] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d48143-869a-9f6a-e853-dfb354026c18, 'name': SearchDatastore_Task, 'duration_secs': 0.010857} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.668908] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 925.669211] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 925.669405] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.669554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.669737] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 925.670023] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-467ac414-fcab-4ea9-97e0-99f3845ee3cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.678062] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 925.678252] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 925.679051] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c0760040-8ab1-48d4-88fe-1b0f2195d05f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.684453] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 925.684453] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5241fb41-cae3-14fa-d245-ad0ee6b655dd" [ 925.684453] env[62000]: _type = "Task" [ 925.684453] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.694184] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5241fb41-cae3-14fa-d245-ad0ee6b655dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.700068] env[62000]: DEBUG nova.network.neutron [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updated VIF entry in instance network info cache for port e041eaad-7b03-44b1-be56-0c3871c66ecc. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 925.700543] env[62000]: DEBUG nova.network.neutron [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updating instance_info_cache with network_info: [{"id": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "address": "fa:16:3e:4c:38:15", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape041eaad-7b", "ovs_interfaceid": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.023852] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 926.031705] env[62000]: INFO nova.compute.claims [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 926.059517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9aab2c8b-85b3-4b93-beed-106593727f75 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.161s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.117354] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882579, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.186380] env[62000]: DEBUG nova.network.neutron [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Updating instance_info_cache with network_info: [{"id": "5b734604-c3c4-4667-851c-a635f7aeafc6", "address": "fa:16:3e:fc:98:c7", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b734604-c3", "ovs_interfaceid": "5b734604-c3c4-4667-851c-a635f7aeafc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.200468] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5241fb41-cae3-14fa-d245-ad0ee6b655dd, 'name': SearchDatastore_Task, 'duration_secs': 0.011652} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.202396] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c677db01-38fc-48c5-8285-c0c53f699cae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.205939] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90fabb7-28ab-445d-aceb-6e92483977b6 req-deb1af13-30d7-4750-b6d0-f5ce10577dcb service nova] Releasing lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.210809] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 926.210809] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5219868c-a928-fe93-f61f-2faa4d77bef5" [ 926.210809] env[62000]: _type = "Task" [ 926.210809] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.221208] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5219868c-a928-fe93-f61f-2faa4d77bef5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.412790] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "bf606000-346b-48db-972a-000b54a8ec5d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.413169] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.413528] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "bf606000-346b-48db-972a-000b54a8ec5d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.413775] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.413932] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.416377] env[62000]: INFO nova.compute.manager [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Terminating instance [ 926.418223] env[62000]: DEBUG nova.compute.manager [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.418422] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 926.419356] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbeaeea7-015c-42c1-8f3e-36d353ee073e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.428060] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 926.428320] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d48e90ce-1420-41cc-baa8-bea00bdaee06 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.434825] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 926.434825] env[62000]: value = "task-882580" [ 926.434825] env[62000]: _type = "Task" [ 926.434825] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.443332] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882580, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.476412] env[62000]: DEBUG nova.compute.manager [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.476592] env[62000]: DEBUG oslo_concurrency.lockutils [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.476821] env[62000]: DEBUG oslo_concurrency.lockutils [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.477040] env[62000]: DEBUG oslo_concurrency.lockutils [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.477234] env[62000]: DEBUG nova.compute.manager [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] No waiting events found dispatching network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.477427] env[62000]: WARNING nova.compute.manager [req-fe97fbb1-2c78-4ada-8c9b-34b67f939717 req-b9bad855-6aba-4020-bff2-2f1b0a2b4302 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received unexpected event network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f for instance with vm_state active and task_state None. [ 926.527445] env[62000]: DEBUG nova.network.neutron [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Successfully updated port: 68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 926.538829] env[62000]: INFO nova.compute.resource_tracker [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating resource usage from migration 3ce7c346-1ce0-403a-9471-e10958a315c3 [ 926.613457] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882579, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.584761} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.616540] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 926.616799] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 926.617355] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-09de6137-775a-4514-8940-80469a94956c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.625459] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 926.625459] env[62000]: value = "task-882581" [ 926.625459] env[62000]: _type = "Task" [ 926.625459] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.639327] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882581, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.694145] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.695149] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Instance network_info: |[{"id": "5b734604-c3c4-4667-851c-a635f7aeafc6", "address": "fa:16:3e:fc:98:c7", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b734604-c3", "ovs_interfaceid": "5b734604-c3c4-4667-851c-a635f7aeafc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 926.695668] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:98:c7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5b734604-c3c4-4667-851c-a635f7aeafc6', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 926.705146] env[62000]: DEBUG oslo.service.loopingcall [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 926.708941] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 926.709619] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a678992e-fb21-4251-a444-a4467ee95462 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.739046] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5219868c-a928-fe93-f61f-2faa4d77bef5, 'name': SearchDatastore_Task, 'duration_secs': 0.010754} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.742581] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.742891] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 0cbc29c0-b817-4960-9a8d-5fef0b439d58/0cbc29c0-b817-4960-9a8d-5fef0b439d58.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 926.743150] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 926.743150] env[62000]: value = "task-882582" [ 926.743150] env[62000]: _type = "Task" [ 926.743150] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.744022] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-52267409-1319-4f20-bbb2-0ea8d6ae69ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.754366] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882582, 'name': CreateVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.758095] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 926.758095] env[62000]: value = "task-882583" [ 926.758095] env[62000]: _type = "Task" [ 926.758095] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.766760] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882583, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.805950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ac89546-6071-4f48-bdeb-bf5d2a8fd4fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.814838] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bcd97c1-7c1c-44b7-a5e3-a103de66389c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.845829] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9c80c4-757b-4c90-b3af-ebac1672b90f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.854391] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e820ccb-5db2-4d3b-977e-3d30bbb7d654 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.867996] env[62000]: DEBUG nova.compute.provider_tree [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.946290] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882580, 'name': PowerOffVM_Task, 'duration_secs': 0.241938} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.946699] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 926.946976] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 926.947337] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9a2405e-2c9c-49dc-8e05-33ace30c31eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.993658] env[62000]: DEBUG nova.compute.manager [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Received event network-vif-plugged-5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.993946] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Acquiring lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.994188] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.994389] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.994581] env[62000]: DEBUG nova.compute.manager [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] No waiting events found dispatching network-vif-plugged-5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 926.994752] env[62000]: WARNING nova.compute.manager [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Received unexpected event network-vif-plugged-5b734604-c3c4-4667-851c-a635f7aeafc6 for instance with vm_state building and task_state spawning. [ 926.994917] env[62000]: DEBUG nova.compute.manager [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Received event network-changed-5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 926.995245] env[62000]: DEBUG nova.compute.manager [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Refreshing instance network info cache due to event network-changed-5b734604-c3c4-4667-851c-a635f7aeafc6. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 926.995466] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Acquiring lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 926.995611] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Acquired lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 926.995777] env[62000]: DEBUG nova.network.neutron [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Refreshing network info cache for port 5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 927.012807] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 927.013044] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 927.013242] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore2] bf606000-346b-48db-972a-000b54a8ec5d {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.013775] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5c8252c2-3fbc-423b-85b4-3bdc2e1aae3e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.021254] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 927.021254] env[62000]: value = "task-882585" [ 927.021254] env[62000]: _type = "Task" [ 927.021254] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.029114] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882585, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.033091] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.033310] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.033536] env[62000]: DEBUG nova.network.neutron [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.135540] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882581, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.151509} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.135906] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 927.136632] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7481d52-d20c-4182-bf83-32bfe1953332 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.158628] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 927.158978] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c4c734b-9d78-4e56-ba48-d78a83de2109 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.180960] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 927.180960] env[62000]: value = "task-882586" [ 927.180960] env[62000]: _type = "Task" [ 927.180960] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.190145] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882586, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.254993] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882582, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.267200] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882583, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.289261] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.289733] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.290076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.290358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.290605] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.293181] env[62000]: INFO nova.compute.manager [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Terminating instance [ 927.295065] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "refresh_cache-0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.295239] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquired lock "refresh_cache-0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.295422] env[62000]: DEBUG nova.network.neutron [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 927.371073] env[62000]: DEBUG nova.scheduler.client.report [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.531296] env[62000]: DEBUG oslo_vmware.api [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882585, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.377889} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.531693] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.531791] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 927.531980] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 927.532239] env[62000]: INFO nova.compute.manager [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Took 1.11 seconds to destroy the instance on the hypervisor. [ 927.532496] env[62000]: DEBUG oslo.service.loopingcall [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.532691] env[62000]: DEBUG nova.compute.manager [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.532787] env[62000]: DEBUG nova.network.neutron [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 927.581785] env[62000]: WARNING nova.network.neutron [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] e21c3795-1c6b-42ef-af81-e113912fa80d already exists in list: networks containing: ['e21c3795-1c6b-42ef-af81-e113912fa80d']. ignoring it [ 927.695733] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882586, 'name': ReconfigVM_Task, 'duration_secs': 0.461783} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.696143] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 927.699498] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9051554a-7949-4cc3-ac58-93dd038ad042 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.707291] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 927.707291] env[62000]: value = "task-882587" [ 927.707291] env[62000]: _type = "Task" [ 927.707291] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.715708] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882587, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.755843] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882582, 'name': CreateVM_Task, 'duration_secs': 0.576374} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.756984] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 927.758363] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.758363] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.758604] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 927.758973] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c04439f-7484-4f41-b594-9cd8f1303e44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.766843] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 927.766843] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ea18b8-7178-2187-7f74-ac151d02b9ce" [ 927.766843] env[62000]: _type = "Task" [ 927.766843] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.770825] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882583, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.725121} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.774027] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 0cbc29c0-b817-4960-9a8d-5fef0b439d58/0cbc29c0-b817-4960-9a8d-5fef0b439d58.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 927.774414] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 927.776827] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b7e28f1e-7bdf-415d-9967-6abacec7ab02 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.785435] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 927.785755] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Processing image d19f59f0-69e8-45e4-9559-cc7320fda017 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 927.786076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 927.787622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 927.787622] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 927.787622] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 927.787622] env[62000]: value = "task-882588" [ 927.787622] env[62000]: _type = "Task" [ 927.787622] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.790208] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d058fc96-5037-425f-b250-758dadc31c3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.803642] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.810016] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 927.810886] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 927.811747] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29eab320-c13b-488d-9cc5-afbe9a223dcb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.817812] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 927.817812] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5239ebac-7eb2-0152-554b-bdffb589cbf8" [ 927.817812] env[62000]: _type = "Task" [ 927.817812] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.818658] env[62000]: DEBUG nova.network.neutron [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Updated VIF entry in instance network info cache for port 5b734604-c3c4-4667-851c-a635f7aeafc6. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 927.819150] env[62000]: DEBUG nova.network.neutron [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Updating instance_info_cache with network_info: [{"id": "5b734604-c3c4-4667-851c-a635f7aeafc6", "address": "fa:16:3e:fc:98:c7", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5b734604-c3", "ovs_interfaceid": "5b734604-c3c4-4667-851c-a635f7aeafc6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 927.829226] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5239ebac-7eb2-0152-554b-bdffb589cbf8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.831086] env[62000]: DEBUG nova.network.neutron [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 927.876386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.353s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.876656] env[62000]: INFO nova.compute.manager [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Migrating [ 927.884622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.466s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.885118] env[62000]: DEBUG nova.objects.instance [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'pci_requests' on Instance uuid 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.125413] env[62000]: DEBUG nova.network.neutron [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.217623] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882587, 'name': Rename_Task, 'duration_secs': 0.189373} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.217975] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 928.218263] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-399e9433-63d1-4805-9d48-28532d5e351b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.226032] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 928.226032] env[62000]: value = "task-882589" [ 928.226032] env[62000]: _type = "Task" [ 928.226032] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.235831] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882589, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.273636] env[62000]: DEBUG nova.network.neutron [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "68fd0d34-536e-4504-a837-62eca9533e5f", "address": "fa:16:3e:01:88:92", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fd0d34-53", "ovs_interfaceid": "68fd0d34-536e-4504-a837-62eca9533e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.300565] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.10522} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.300848] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 928.301712] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35db8d52-38f4-480e-95a5-91d5e14fb481 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.323783] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 0cbc29c0-b817-4960-9a8d-5fef0b439d58/0cbc29c0-b817-4960-9a8d-5fef0b439d58.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 928.324647] env[62000]: DEBUG oslo_concurrency.lockutils [req-a9c73895-5d59-4031-ade6-8a9eee03bc2d req-a1e70af8-15a1-40e1-9e73-4c974a93cbc2 service nova] Releasing lock "refresh_cache-54eb64ad-bc7b-4495-9e93-451059139c4c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.325064] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd05c5f6-d69f-49bd-b305-eae897a5d0d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.348935] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 928.349195] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Fetch image to [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35/OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 928.349391] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Downloading stream optimized image d19f59f0-69e8-45e4-9559-cc7320fda017 to [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35/OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35.vmdk on the data store datastore1 as vApp {{(pid=62000) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 928.349567] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Downloading image file data d19f59f0-69e8-45e4-9559-cc7320fda017 to the ESX as VM named 'OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35' {{(pid=62000) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 928.351485] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 928.351485] env[62000]: value = "task-882590" [ 928.351485] env[62000]: _type = "Task" [ 928.351485] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.359908] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882590, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.403607] env[62000]: DEBUG nova.objects.instance [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'numa_topology' on Instance uuid 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.404986] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.405818] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.405818] env[62000]: DEBUG nova.network.neutron [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 928.435047] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 928.435047] env[62000]: value = "resgroup-9" [ 928.435047] env[62000]: _type = "ResourcePool" [ 928.435047] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 928.435350] env[62000]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-5dc016c2-4fba-4095-8fb0-89f553fb03d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.456441] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease: (returnval){ [ 928.456441] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 928.456441] env[62000]: _type = "HttpNfcLease" [ 928.456441] env[62000]: } obtained for vApp import into resource pool (val){ [ 928.456441] env[62000]: value = "resgroup-9" [ 928.456441] env[62000]: _type = "ResourcePool" [ 928.456441] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 928.456797] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the lease: (returnval){ [ 928.456797] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 928.456797] env[62000]: _type = "HttpNfcLease" [ 928.456797] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 928.464047] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.464047] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 928.464047] env[62000]: _type = "HttpNfcLease" [ 928.464047] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 928.613513] env[62000]: DEBUG nova.compute.manager [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 928.613784] env[62000]: DEBUG nova.compute.manager [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-68fd0d34-536e-4504-a837-62eca9533e5f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 928.614024] env[62000]: DEBUG oslo_concurrency.lockutils [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.628506] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Releasing lock "refresh_cache-0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.628867] env[62000]: DEBUG nova.compute.manager [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 928.629084] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 928.630178] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7849b96b-0c5c-4164-8536-571d55cbd046 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.637409] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 928.637678] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20b32db7-117c-4627-91bf-f593474844f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.644057] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 928.644057] env[62000]: value = "task-882592" [ 928.644057] env[62000]: _type = "Task" [ 928.644057] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.651462] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882592, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.683769] env[62000]: DEBUG nova.network.neutron [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.736584] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882589, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.777557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 928.777557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 928.777557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.777837] env[62000]: DEBUG oslo_concurrency.lockutils [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 928.778031] env[62000]: DEBUG nova.network.neutron [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port 68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 928.779822] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91e709f6-5af1-433d-91c0-c299ab5623f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.799421] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 928.799699] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 928.799813] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 928.800017] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 928.800198] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 928.800350] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 928.800558] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 928.800721] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 928.800897] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 928.801071] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 928.801255] env[62000]: DEBUG nova.virt.hardware [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 928.807636] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfiguring VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 928.808704] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9e37225-cc8d-4b16-bdee-a24c94417efa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.826535] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 928.826535] env[62000]: value = "task-882593" [ 928.826535] env[62000]: _type = "Task" [ 928.826535] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.835713] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882593, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.861905] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.910125] env[62000]: INFO nova.compute.claims [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 928.965686] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 928.965686] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 928.965686] env[62000]: _type = "HttpNfcLease" [ 928.965686] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 929.125107] env[62000]: DEBUG nova.compute.manager [req-724416e5-5f22-4300-a3b2-318f945bbda8 req-04799674-ace0-464a-8e53-99a7e8d2dbd4 service nova] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Received event network-vif-deleted-060c8648-df77-440f-81c1-8d8d84a20a5d {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 929.133423] env[62000]: DEBUG nova.network.neutron [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.154027] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882592, 'name': PowerOffVM_Task, 'duration_secs': 0.137779} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.154398] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 929.154579] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 929.154822] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1908b8fa-a81b-4cf8-ac4d-b365ea83c770 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.181431] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 929.181800] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 929.181914] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleting the datastore file [datastore2] 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 929.182224] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-19b45cf6-f2ed-49b2-9f99-5e071688e06f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.186206] env[62000]: INFO nova.compute.manager [-] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Took 1.65 seconds to deallocate network for instance. [ 929.191684] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for the task: (returnval){ [ 929.191684] env[62000]: value = "task-882595" [ 929.191684] env[62000]: _type = "Task" [ 929.191684] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.201238] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882595, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.236147] env[62000]: DEBUG oslo_vmware.api [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882589, 'name': PowerOnVM_Task, 'duration_secs': 0.782431} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.236477] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 929.236695] env[62000]: INFO nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Took 9.32 seconds to spawn the instance on the hypervisor. [ 929.236880] env[62000]: DEBUG nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 929.237722] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e993be49-a3d0-44a4-8b57-15e313cc29a1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.336559] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882593, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.364104] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882590, 'name': ReconfigVM_Task, 'duration_secs': 0.625163} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.366636] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 0cbc29c0-b817-4960-9a8d-5fef0b439d58/0cbc29c0-b817-4960-9a8d-5fef0b439d58.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 929.367323] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5516ee35-ba3c-4508-8317-5e5e2b5a2bcd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.374457] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 929.374457] env[62000]: value = "task-882596" [ 929.374457] env[62000]: _type = "Task" [ 929.374457] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.382512] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882596, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.467265] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 929.467265] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 929.467265] env[62000]: _type = "HttpNfcLease" [ 929.467265] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 929.467560] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 929.467560] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fc8ce9-8e96-176d-0659-d6d842886db8" [ 929.467560] env[62000]: _type = "HttpNfcLease" [ 929.467560] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 929.468261] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd44349-d9d7-4f44-8ee2-917a44d29710 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.477485] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 929.477694] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 929.535752] env[62000]: DEBUG nova.network.neutron [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port 68fd0d34-536e-4504-a837-62eca9533e5f. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 929.536216] env[62000]: DEBUG nova.network.neutron [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "68fd0d34-536e-4504-a837-62eca9533e5f", "address": "fa:16:3e:01:88:92", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fd0d34-53", "ovs_interfaceid": "68fd0d34-536e-4504-a837-62eca9533e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.544162] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-930931d3-c3f5-4289-bcc2-848c43006843 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.636862] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 929.693505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.702422] env[62000]: DEBUG oslo_vmware.api [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Task: {'id': task-882595, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106525} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.702671] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.702859] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 929.703079] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 929.703282] env[62000]: INFO nova.compute.manager [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Took 1.07 seconds to destroy the instance on the hypervisor. [ 929.703523] env[62000]: DEBUG oslo.service.loopingcall [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.703712] env[62000]: DEBUG nova.compute.manager [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.703806] env[62000]: DEBUG nova.network.neutron [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 929.718957] env[62000]: DEBUG nova.network.neutron [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 929.756272] env[62000]: INFO nova.compute.manager [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Took 22.82 seconds to build instance. [ 929.837363] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882593, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.883231] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882596, 'name': Rename_Task, 'duration_secs': 0.257215} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.883504] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 929.883745] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-394b7038-996f-4419-9f8f-4c069a3cf3b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.889797] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 929.889797] env[62000]: value = "task-882597" [ 929.889797] env[62000]: _type = "Task" [ 929.889797] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.897136] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882597, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.039926] env[62000]: DEBUG oslo_concurrency.lockutils [req-21fe5ba6-a78a-4262-b390-ae6dbff598e3 req-b193f3ec-5fe8-4603-a85c-fdc4fc875cf4 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.171824] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a965c33-30d9-48b9-925c-923c3157d987 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.182063] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6eb4e5c-ed55-43d6-8337-06acef95eb9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.219435] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6309d76-3fe6-482d-8d92-452aee87c734 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.222444] env[62000]: DEBUG nova.network.neutron [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.233167] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874ad627-7155-417f-99a2-2a52e4c7e4a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.250404] env[62000]: DEBUG nova.compute.provider_tree [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.258858] env[62000]: DEBUG oslo_concurrency.lockutils [None req-7b8bf819-d539-464b-a6d4-0a2490c0e71c tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.336s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.339493] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882593, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.407134] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882597, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.548949] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 930.549254] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 930.550565] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fe3b55-f168-47c7-a545-a028b159c033 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.557916] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 930.558133] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 930.558387] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-64a6ae23-7b49-4e7d-8661-420cea6080c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.726416] env[62000]: INFO nova.compute.manager [-] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Took 1.02 seconds to deallocate network for instance. [ 930.755610] env[62000]: DEBUG nova.scheduler.client.report [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 930.837841] env[62000]: DEBUG oslo_vmware.api [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882593, 'name': ReconfigVM_Task, 'duration_secs': 1.623833} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.838733] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.838904] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfigured VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 930.901312] env[62000]: DEBUG oslo_vmware.api [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882597, 'name': PowerOnVM_Task, 'duration_secs': 0.711694} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.901312] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 930.901312] env[62000]: INFO nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Took 8.61 seconds to spawn the instance on the hypervisor. [ 930.901870] env[62000]: DEBUG nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.902425] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed043c2e-bb78-4baa-a2d1-f4144a0700a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.915815] env[62000]: DEBUG nova.compute.manager [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Received event network-changed-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 930.915815] env[62000]: DEBUG nova.compute.manager [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Refreshing instance network info cache due to event network-changed-1eb9c5c5-8fe1-4652-986f-3b9abdd36485. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 930.915815] env[62000]: DEBUG oslo_concurrency.lockutils [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.915815] env[62000]: DEBUG oslo_concurrency.lockutils [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.915815] env[62000]: DEBUG nova.network.neutron [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Refreshing network info cache for port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 930.955399] env[62000]: DEBUG oslo_vmware.rw_handles [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5228c667-9560-f839-2a18-e52a6806d7f4/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 930.955652] env[62000]: INFO nova.virt.vmwareapi.images [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Downloaded image file data d19f59f0-69e8-45e4-9559-cc7320fda017 [ 930.956739] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f017dc-d86c-4f28-9bf8-26548c6b35d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.973113] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3b5cbe1f-fb1d-4df1-b280-2d34ecf23f68 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.095715] env[62000]: INFO nova.virt.vmwareapi.images [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] The imported VM was unregistered [ 931.098358] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 931.098358] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 931.098787] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-05d7b9f5-d203-49da-8c39-176617e3535a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.109746] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 931.110472] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35/OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35.vmdk to [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk. {{(pid=62000) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 931.110472] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-5422e726-80cf-40c0-9458-1670f2b87890 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.117212] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 931.117212] env[62000]: value = "task-882599" [ 931.117212] env[62000]: _type = "Task" [ 931.117212] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.125529] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.156468] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7011da8-b710-4d76-a155-b118c2a750fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.175374] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 931.233262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.261451] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.377s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.264365] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.571s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.264751] env[62000]: DEBUG nova.objects.instance [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid bf606000-346b-48db-972a-000b54a8ec5d {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.301430] env[62000]: INFO nova.network.neutron [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 931.344124] env[62000]: DEBUG oslo_concurrency.lockutils [None req-81ecbe99-d3f5-4229-9e0a-f1379754ce77 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.033s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.424131] env[62000]: INFO nova.compute.manager [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Took 22.21 seconds to build instance. [ 931.525385] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.525716] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.607932] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.608653] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.628988] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.682124] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 931.682850] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1fcde924-0455-4e58-9565-e59888262c2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.690629] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 931.690629] env[62000]: value = "task-882600" [ 931.690629] env[62000]: _type = "Task" [ 931.690629] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.703484] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882600, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.780936] env[62000]: DEBUG nova.network.neutron [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updated VIF entry in instance network info cache for port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 931.781485] env[62000]: DEBUG nova.network.neutron [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.926096] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1e573c36-8910-4657-b8d2-19e23d3ca60e tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.718s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.010184] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30304a71-9741-4737-a360-981073e6c824 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.018342] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-faf9825f-5b8a-4bd7-ae4e-0bec60acc673 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.049029] env[62000]: INFO nova.compute.manager [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Detaching volume acf81f65-f409-4b80-9f5c-10c061b9ad88 [ 932.052311] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d75eb6-8adf-4fc4-bad1-0579127c3d27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.063251] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c93579e-bf09-4203-ba05-04c418d28dca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.077497] env[62000]: DEBUG nova.compute.provider_tree [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.085046] env[62000]: INFO nova.virt.block_device [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Attempting to driver detach volume acf81f65-f409-4b80-9f5c-10c061b9ad88 from mountpoint /dev/sdb [ 932.085046] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 932.085258] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201587', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'name': 'volume-acf81f65-f409-4b80-9f5c-10c061b9ad88', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66edb906-f936-4f57-833d-224f36af109e', 'attached_at': '', 'detached_at': '', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'serial': 'acf81f65-f409-4b80-9f5c-10c061b9ad88'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 932.086444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36d01ff6-7cfb-4fcd-90f5-c66d36d417a1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.108740] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ec67a8-cb4f-435b-8ca6-08a00c3e6523 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.112301] env[62000]: DEBUG nova.compute.utils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.119203] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3387cb79-90df-43a3-9f6e-4ca325da7334 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.149610] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5121adda-eba1-4727-9764-962fa2555785 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.152528] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.166564] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] The volume has not been displaced from its original location: [datastore1] volume-acf81f65-f409-4b80-9f5c-10c061b9ad88/volume-acf81f65-f409-4b80-9f5c-10c061b9ad88.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 932.172231] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfiguring VM instance instance-00000040 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 932.172686] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1b13932-3e1c-464c-9161-2fa646efb9ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.192236] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 932.192236] env[62000]: value = "task-882601" [ 932.192236] env[62000]: _type = "Task" [ 932.192236] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.204105] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882601, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.207268] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882600, 'name': PowerOffVM_Task, 'duration_secs': 0.26247} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.207536] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 932.207727] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 932.284285] env[62000]: DEBUG oslo_concurrency.lockutils [req-af99d29e-5609-4470-99b2-4afdbd16df73 req-afbd7af0-9c00-4f2d-81a1-77b3a2aae7c7 service nova] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.581178] env[62000]: DEBUG nova.scheduler.client.report [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 932.616257] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.631796] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.705747] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882601, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.714999] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 932.715294] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 932.715553] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 932.715778] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 932.715935] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 932.716109] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 932.716372] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 932.716544] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 932.716718] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 932.716891] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 932.717086] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 932.722400] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-20316837-c651-45ec-b5c1-ac788b374326 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.738846] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 932.738846] env[62000]: value = "task-882602" [ 932.738846] env[62000]: _type = "Task" [ 932.738846] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.748027] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882602, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.052408] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.052793] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.052904] env[62000]: DEBUG nova.network.neutron [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 933.086511] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.822s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.089933] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.857s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.090155] env[62000]: DEBUG nova.objects.instance [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lazy-loading 'resources' on Instance uuid 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 933.116628] env[62000]: INFO nova.scheduler.client.report [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance bf606000-346b-48db-972a-000b54a8ec5d [ 933.134616] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.208954] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882601, 'name': ReconfigVM_Task, 'duration_secs': 0.554658} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.209290] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Reconfigured VM instance instance-00000040 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 933.215204] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4a3f8ca-7ac1-49f7-91f4-033f3a9d95d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.232542] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 933.232542] env[62000]: value = "task-882603" [ 933.232542] env[62000]: _type = "Task" [ 933.232542] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.243121] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882603, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.251889] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882602, 'name': ReconfigVM_Task, 'duration_secs': 0.265632} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.252433] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 933.397418] env[62000]: DEBUG nova.compute.manager [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 933.398396] env[62000]: DEBUG oslo_concurrency.lockutils [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.399040] env[62000]: DEBUG oslo_concurrency.lockutils [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.400173] env[62000]: DEBUG oslo_concurrency.lockutils [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.400173] env[62000]: DEBUG nova.compute.manager [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] No waiting events found dispatching network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.400173] env[62000]: WARNING nova.compute.manager [req-9ccc2baf-4303-4c2b-af4c-9257e98427d4 req-eae51cda-7d2a-4b4a-a25d-92e2c67d5555 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received unexpected event network-vif-plugged-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 for instance with vm_state shelved_offloaded and task_state spawning. [ 933.423129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.423449] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.423684] env[62000]: INFO nova.compute.manager [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Shelving [ 933.629838] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2a563f75-7b76-475f-99b1-b5766929700c tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "bf606000-346b-48db-972a-000b54a8ec5d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.217s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.641662] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.699836] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.700111] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.700351] env[62000]: INFO nova.compute.manager [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Attaching volume 42495829-8539-4f47-9949-eb04eac3dc16 to /dev/sdb [ 933.741360] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0c1944-65e4-49e6-8f00-faf703ebedf8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.751390] env[62000]: DEBUG oslo_vmware.api [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882603, 'name': ReconfigVM_Task, 'duration_secs': 0.178463} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.753388] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201587', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'name': 'volume-acf81f65-f409-4b80-9f5c-10c061b9ad88', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '66edb906-f936-4f57-833d-224f36af109e', 'attached_at': '', 'detached_at': '', 'volume_id': 'acf81f65-f409-4b80-9f5c-10c061b9ad88', 'serial': 'acf81f65-f409-4b80-9f5c-10c061b9ad88'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 933.756499] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4415763a-9d13-4ce4-b073-ce3dea541621 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.762920] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 933.763214] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 933.763380] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 933.763572] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 933.763721] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 933.763867] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 933.764194] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 933.764270] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 933.764398] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 933.764559] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 933.764734] env[62000]: DEBUG nova.virt.hardware [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 933.769935] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfiguring VM instance instance-00000054 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 933.770419] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-119c6115-b6bc-4d64-9c30-afcdb7b76fa8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.803256] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 933.803256] env[62000]: value = "task-882604" [ 933.803256] env[62000]: _type = "Task" [ 933.803256] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.806861] env[62000]: DEBUG nova.virt.block_device [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating existing volume attachment record: 46e7272b-4202-4d1c-8dea-5986516150b8 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 933.821796] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882604, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.872514] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edd0534d-4c7e-4c98-86ba-8fc2bb83455d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.876928] env[62000]: DEBUG nova.network.neutron [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.882717] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d10945f-720c-4f18-aa80-dfac1cab3794 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.913610] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc19ada9-6ce1-47db-8d97-831e48cee5c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.921151] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50668908-0483-4954-b7d7-a7fb3bc79c3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.936128] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.936457] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.939095] env[62000]: DEBUG nova.compute.provider_tree [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 933.942336] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 933.942599] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-735b0f29-86aa-4502-bb2f-4d546b79e73e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.949047] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 933.949047] env[62000]: value = "task-882606" [ 933.949047] env[62000]: _type = "Task" [ 933.949047] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.957050] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882606, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.134912] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882599, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.696554} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.135635] env[62000]: INFO nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35/OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35.vmdk to [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk. [ 934.135635] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Cleaning up location [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 934.135757] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_df8f2c34-979f-4cb3-b51a-0e23e8262f35 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 934.136037] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e6d9ec02-d055-4357-a11c-fe15a2bbd3ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.142289] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 934.142289] env[62000]: value = "task-882609" [ 934.142289] env[62000]: _type = "Task" [ 934.142289] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.150400] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882609, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.298312] env[62000]: DEBUG nova.objects.instance [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 934.318050] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882604, 'name': ReconfigVM_Task, 'duration_secs': 0.389004} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.318805] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfigured VM instance instance-00000054 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 934.319910] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917fe12e-d134-4842-8f71-984d3dfc3363 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.343227] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.343839] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f2a534e9-a8db-4be9-bddf-496da25d7124 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.364826] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 934.364826] env[62000]: value = "task-882610" [ 934.364826] env[62000]: _type = "Task" [ 934.364826] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.373704] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882610, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.379603] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.408614] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='0a92766c12701b84f26ca5cf11da296d',container_format='bare',created_at=2024-09-16T06:59:01Z,direct_url=,disk_format='vmdk',id=fec7d673-ed56-4ee2-8361-5001b5b340d5,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1132952803-shelved',owner='cac5f0a5704d434082131155e107d190',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-09-16T06:59:18Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.408614] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.408809] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.408854] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.409517] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.409517] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.409517] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.409517] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.409698] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.409821] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.410031] env[62000]: DEBUG nova.virt.hardware [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.411592] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a547bbe-1fd5-4c9c-83ba-f174cc82d60c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.419641] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d86ebba6-8a00-446c-9066-0d0dd21a668c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.433317] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6c:7d:d5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 934.440912] env[62000]: DEBUG oslo.service.loopingcall [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 934.441295] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.441459] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.441695] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 934.442437] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94367e96-0e1e-4b08-9370-8ab3a7c21919 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.444960] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-57ddf477-6c95-4445-bc2e-ebec9e5965b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.459787] env[62000]: DEBUG nova.scheduler.client.report [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 934.487021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b8bb326-1e50-441d-b909-ea9d8b3bc37e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.487021] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 934.487021] env[62000]: value = "task-882611" [ 934.487021] env[62000]: _type = "Task" [ 934.487021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.490319] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882606, 'name': PowerOffVM_Task, 'duration_secs': 0.312454} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.215546] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 935.234549] env[62000]: DEBUG oslo_concurrency.lockutils [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.235196] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.146s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.237731] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffe53224-9e73-4e82-ace5-376be50107ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.245413] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfiguring VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 935.257673] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-28abf1a2-cb41-4a61-b2b4-84a2d02eaafe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.271226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.271438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.273559] env[62000]: INFO nova.scheduler.client.report [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Deleted allocations for instance 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b [ 935.278943] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882611, 'name': CreateVM_Task, 'duration_secs': 0.562353} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.282105] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 935.296393] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.296572] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.296957] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 935.304415] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43c65e53-3b1c-4a8c-9a42-fa060fcff60d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.306617] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2cf0370-7dc6-4da8-8773-c36bc1cd3f83 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.309939] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 935.309939] env[62000]: value = "task-882612" [ 935.309939] env[62000]: _type = "Task" [ 935.309939] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.310218] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882610, 'name': ReconfigVM_Task, 'duration_secs': 0.804613} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.310708] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882609, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.065738} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.311093] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Reconfigured VM instance instance-00000054 to attach disk [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225/dfa26fab-40ff-4409-bb73-79aa555ab225.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.311409] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 935.314879] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.315095] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.315385] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk to [datastore1] 54eb64ad-bc7b-4495-9e93-451059139c4c/54eb64ad-bc7b-4495-9e93-451059139c4c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 935.320688] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-515ccf9d-3c9f-407e-b234-8630e97eec94 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.327787] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 935.327787] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]528e1a33-b9fb-1544-3e69-2529740ab214" [ 935.327787] env[62000]: _type = "Task" [ 935.327787] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.343412] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.343903] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 935.343903] env[62000]: value = "task-882613" [ 935.343903] env[62000]: _type = "Task" [ 935.343903] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.354176] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.354594] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Processing image fec7d673-ed56-4ee2-8361-5001b5b340d5 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.354952] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.355249] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.355779] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.362935] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b659cae9-05c7-48b3-8ce1-b235cb735278 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.367974] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.382241] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.382761] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 935.383671] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b2b03f3-f60b-4e43-9d8f-2e1e5d14b494 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.389250] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 935.389250] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52668600-df68-dc99-d4e9-08521455f746" [ 935.389250] env[62000]: _type = "Task" [ 935.389250] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.396672] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52668600-df68-dc99-d4e9-08521455f746, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.427174] env[62000]: DEBUG nova.compute.manager [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 935.427503] env[62000]: DEBUG nova.compute.manager [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing instance network info cache due to event network-changed-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 935.427765] env[62000]: DEBUG oslo_concurrency.lockutils [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] Acquiring lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.427964] env[62000]: DEBUG oslo_concurrency.lockutils [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] Acquired lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.428211] env[62000]: DEBUG nova.network.neutron [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Refreshing network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 935.773873] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d8ca1c2c-6b06-420d-a5da-7bf15713fe9a tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.248s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.775156] env[62000]: DEBUG oslo_concurrency.lockutils [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.541s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.775405] env[62000]: DEBUG nova.compute.manager [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 935.776294] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b331672-ef27-443d-b7f4-45caf07a2e84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.779834] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 935.787250] env[62000]: DEBUG nova.compute.manager [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 935.787819] env[62000]: DEBUG nova.objects.instance [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.789904] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2bb79bda-0a47-45c9-9711-7533d0c61310 tempest-ServerShowV247Test-2062924323 tempest-ServerShowV247Test-2062924323-project-member] Lock "0f5b416a-20e4-42f5-9ad9-a8489ab11e3b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.500s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.819031] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1f3a2f-debb-4db2-9829-3a5e3d588f04 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.828557] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.843210] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 935.843688] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-9a796f07-6add-49a7-86f8-7ac1669bc6a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.846503] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3caa8e2-582a-45af-87d8-0a4a21d8651b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.868610] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 935.876770] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 935.876770] env[62000]: value = "task-882614" [ 935.876770] env[62000]: _type = "Task" [ 935.876770] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.877102] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.885904] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882614, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.901269] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 935.901269] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Fetch image to [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0/OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 935.901269] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Downloading stream optimized image fec7d673-ed56-4ee2-8361-5001b5b340d5 to [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0/OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0.vmdk on the data store datastore2 as vApp {{(pid=62000) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 935.901449] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Downloading image file data fec7d673-ed56-4ee2-8361-5001b5b340d5 to the ESX as VM named 'OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0' {{(pid=62000) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 935.987500] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 935.987500] env[62000]: value = "resgroup-9" [ 935.987500] env[62000]: _type = "ResourcePool" [ 935.987500] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 935.987866] env[62000]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-21694aa0-df01-4724-813b-dc88eed859ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.009544] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease: (returnval){ [ 936.009544] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 936.009544] env[62000]: _type = "HttpNfcLease" [ 936.009544] env[62000]: } obtained for vApp import into resource pool (val){ [ 936.009544] env[62000]: value = "resgroup-9" [ 936.009544] env[62000]: _type = "ResourcePool" [ 936.009544] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 936.010138] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the lease: (returnval){ [ 936.010138] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 936.010138] env[62000]: _type = "HttpNfcLease" [ 936.010138] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 936.016345] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 936.016345] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 936.016345] env[62000]: _type = "HttpNfcLease" [ 936.016345] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 936.234656] env[62000]: DEBUG nova.network.neutron [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updated VIF entry in instance network info cache for port 7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 936.235192] env[62000]: DEBUG nova.network.neutron [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [{"id": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "address": "fa:16:3e:6c:7d:d5", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.186", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7a9e2ced-47", "ovs_interfaceid": "7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.296876] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 936.297265] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f647991f-085e-41a8-aa5b-c1d53c055f8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.305308] env[62000]: DEBUG oslo_vmware.api [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 936.305308] env[62000]: value = "task-882617" [ 936.305308] env[62000]: _type = "Task" [ 936.305308] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.314488] env[62000]: DEBUG oslo_vmware.api [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882617, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.315721] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.316024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.317590] env[62000]: INFO nova.compute.claims [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 936.329406] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.359257] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.391758] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882614, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.421059] env[62000]: DEBUG nova.network.neutron [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Port 65be2bd7-e84a-4cc8-b937-6140dc13b2e4 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 936.518037] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 936.518037] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 936.518037] env[62000]: _type = "HttpNfcLease" [ 936.518037] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 936.738575] env[62000]: DEBUG oslo_concurrency.lockutils [req-bede83e2-2c6d-4dd0-a7dd-7db20ac40470 req-674b8b61-9eec-425a-a3fc-89123de3502c service nova] Releasing lock "refresh_cache-8a11689f-fc00-43f8-9215-8d81daa84400" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.815786] env[62000]: DEBUG oslo_vmware.api [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882617, 'name': PowerOffVM_Task, 'duration_secs': 0.339204} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.819319] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 936.819542] env[62000]: DEBUG nova.compute.manager [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 936.820400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8b200f-a85b-4ea5-8944-d7e37649d31b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.829808] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.860243] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.891190] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882614, 'name': CreateSnapshot_Task, 'duration_secs': 0.861729} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.891540] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 936.892373] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bf7eb0-bde5-4e20-b2cd-8aa9b708b908 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.019990] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 937.019990] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 937.019990] env[62000]: _type = "HttpNfcLease" [ 937.019990] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 937.325257] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.338972] env[62000]: DEBUG oslo_concurrency.lockutils [None req-96d3f2c9-f611-4cee-8c37-7d1d1826eaac tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.564s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.363041] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task} progress is 85%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.410707] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 937.414102] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d2bde0ca-bf39-4588-a649-126acc6f52ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.423317] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 937.423317] env[62000]: value = "task-882618" [ 937.423317] env[62000]: _type = "Task" [ 937.423317] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.442854] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882618, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.450390] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.450712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.450922] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.523174] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 937.523174] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 937.523174] env[62000]: _type = "HttpNfcLease" [ 937.523174] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 937.523174] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 937.523174] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5249aa3a-0242-e1f6-69ee-af0486c1c655" [ 937.523174] env[62000]: _type = "HttpNfcLease" [ 937.523174] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 937.523942] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59911ef8-d35b-4a26-9ad2-a7169d9ec8b1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.531649] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 937.531945] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 937.598152] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-d7e6544c-6533-4042-a02a-fcec91b96c58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.633710] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06c9a2b3-2ceb-45bb-bd3a-67a816eb49d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.642461] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c8f9a4-ac84-49bd-a9fa-e11f275b4978 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.674015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85680c0-428b-4a69-8b66-4f4846d0b5ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.681903] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd5769b-5b46-43c7-9d08-9b125368a6a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.695852] env[62000]: DEBUG nova.compute.provider_tree [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.825843] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.860469] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882613, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.393089} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.860724] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/d19f59f0-69e8-45e4-9559-cc7320fda017/d19f59f0-69e8-45e4-9559-cc7320fda017.vmdk to [datastore1] 54eb64ad-bc7b-4495-9e93-451059139c4c/54eb64ad-bc7b-4495-9e93-451059139c4c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 937.861526] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4baafa30-3fc5-41f8-98cd-f7e032b6626b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.882847] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 54eb64ad-bc7b-4495-9e93-451059139c4c/54eb64ad-bc7b-4495-9e93-451059139c4c.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.883528] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b123a24b-b2fc-410f-9cf9-a7da49e5ec42 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.902856] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 937.902856] env[62000]: value = "task-882619" [ 937.902856] env[62000]: _type = "Task" [ 937.902856] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.910144] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882619, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.931893] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882618, 'name': CloneVM_Task} progress is 93%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.202275] env[62000]: DEBUG nova.scheduler.client.report [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 938.329449] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.357886] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 938.358220] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 938.359117] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21030368-bbd0-468e-86fe-531ba4452ba1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.380107] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2882de8-795f-48b6-84d4-1e88b64f135a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.404376] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 938.409043] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a9159118-6d30-4f68-814e-102c69e13c23 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.432991] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 938.432991] env[62000]: value = "task-882620" [ 938.432991] env[62000]: _type = "Task" [ 938.432991] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.436223] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.441781] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882618, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.448251] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882620, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.498205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.498413] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.498709] env[62000]: DEBUG nova.network.neutron [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 938.586505] env[62000]: DEBUG nova.objects.instance [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.711188] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.711745] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 938.827779] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.919616] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882619, 'name': ReconfigVM_Task, 'duration_secs': 0.771048} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.919976] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 54eb64ad-bc7b-4495-9e93-451059139c4c/54eb64ad-bc7b-4495-9e93-451059139c4c.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.920695] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49c4e890-0a6d-4dae-88e9-56e691d996c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.927040] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 938.927040] env[62000]: value = "task-882621" [ 938.927040] env[62000]: _type = "Task" [ 938.927040] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.937814] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882621, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.943720] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882618, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.950056] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882620, 'name': ReconfigVM_Task, 'duration_secs': 0.469959} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.950442] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.955186] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4f455af-5d95-4108-a3b4-8175e8d768ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.971227] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 938.971227] env[62000]: value = "task-882622" [ 938.971227] env[62000]: _type = "Task" [ 938.971227] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.981404] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.098732] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.098962] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.099137] env[62000]: DEBUG nova.network.neutron [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 939.099332] env[62000]: DEBUG nova.objects.instance [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'info_cache' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 939.217547] env[62000]: DEBUG nova.compute.utils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 939.218991] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 939.219173] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 939.280403] env[62000]: DEBUG nova.network.neutron [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 939.307037] env[62000]: DEBUG nova.policy [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 939.329972] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.425427] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 939.425836] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 939.427286] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cc3bd7f-bc8d-4401-9a7a-88ccc5a10ad5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.446370] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882621, 'name': Rename_Task, 'duration_secs': 0.160121} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.451844] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 939.451844] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 939.451844] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 939.451844] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882618, 'name': CloneVM_Task, 'duration_secs': 1.74833} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.452171] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5850c81-6d9d-49e1-ade3-bc0b10b8ff95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.453609] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-300a0405-9ef1-4461-a6af-9ca312fea6fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.455028] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Created linked-clone VM from snapshot [ 939.455766] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a30fe0-bda4-45ef-a221-b0cb704ee3ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.463769] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Uploading image 73acd51d-de92-4df9-a9a6-9faafd098410 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 939.466811] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 939.466811] env[62000]: value = "task-882623" [ 939.466811] env[62000]: _type = "Task" [ 939.466811] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.474117] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882623, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.482077] env[62000]: DEBUG oslo_vmware.api [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882622, 'name': ReconfigVM_Task, 'duration_secs': 0.144796} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.482476] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 939.497804] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 939.497804] env[62000]: value = "vm-201619" [ 939.497804] env[62000]: _type = "VirtualMachine" [ 939.497804] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 939.498148] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-3c712790-7406-496a-b979-e3cbf999f404 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.506209] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lease: (returnval){ [ 939.506209] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523a6cae-52e4-b941-af81-2e2aafad3006" [ 939.506209] env[62000]: _type = "HttpNfcLease" [ 939.506209] env[62000]: } obtained for exporting VM: (result){ [ 939.506209] env[62000]: value = "vm-201619" [ 939.506209] env[62000]: _type = "VirtualMachine" [ 939.506209] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 939.506628] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the lease: (returnval){ [ 939.506628] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523a6cae-52e4-b941-af81-2e2aafad3006" [ 939.506628] env[62000]: _type = "HttpNfcLease" [ 939.506628] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 939.513703] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 939.513703] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523a6cae-52e4-b941-af81-2e2aafad3006" [ 939.513703] env[62000]: _type = "HttpNfcLease" [ 939.513703] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 939.607082] env[62000]: DEBUG nova.objects.base [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Object Instance<66edb906-f936-4f57-833d-224f36af109e> lazy-loaded attributes: flavor,info_cache {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 939.656903] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Successfully created port: 86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 939.724885] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 939.740499] env[62000]: DEBUG oslo_vmware.rw_handles [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5265bae7-7521-303d-ecb3-ac48ddfe01c5/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 939.740499] env[62000]: INFO nova.virt.vmwareapi.images [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Downloaded image file data fec7d673-ed56-4ee2-8361-5001b5b340d5 [ 939.740499] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7867d4a-34e0-4bfb-8126-8cbee60fb8ef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.767369] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cc0a56fe-4831-4f17-9f8d-7bf77385ea27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.784617] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.797401] env[62000]: INFO nova.virt.vmwareapi.images [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] The imported VM was unregistered [ 939.798986] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 939.799486] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 939.799847] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b3796685-02e1-4a3a-9c67-eb3355d0d799 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.813189] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 939.813516] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0/OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0.vmdk to [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk. {{(pid=62000) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 939.813863] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-4ec07d71-aae1-4dab-a6c3-0d5534e43eab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.826121] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 939.826121] env[62000]: value = "task-882626" [ 939.826121] env[62000]: _type = "Task" [ 939.826121] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.829755] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.839915] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.981412] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882623, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.017803] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 940.017803] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523a6cae-52e4-b941-af81-2e2aafad3006" [ 940.017803] env[62000]: _type = "HttpNfcLease" [ 940.017803] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 940.017803] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 940.017803] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523a6cae-52e4-b941-af81-2e2aafad3006" [ 940.017803] env[62000]: _type = "HttpNfcLease" [ 940.017803] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 940.017803] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-627e6f11-a537-4451-9c62-90913f730274 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.025778] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 940.026043] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 940.139657] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ba65c408-615a-4dbf-804c-b243a401ed7d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.317792] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447eaafd-8119-4121-a80a-558b7ece5c87 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.351677] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-046d8ad3-136e-4b97-8f35-aa2ed0405317 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.355332] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.364716] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 940.373784] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.484349] env[62000]: DEBUG oslo_vmware.api [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882623, 'name': PowerOnVM_Task, 'duration_secs': 0.537581} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.484665] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 940.484966] env[62000]: INFO nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Took 15.86 seconds to spawn the instance on the hypervisor. [ 940.485082] env[62000]: DEBUG nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.486396] env[62000]: DEBUG nova.network.neutron [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [{"id": "849930d2-4200-4ec9-ac15-51c243a118bf", "address": "fa:16:3e:07:7f:55", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.229", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap849930d2-42", "ovs_interfaceid": "849930d2-4200-4ec9-ac15-51c243a118bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.488471] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27b4283f-262f-498e-8097-038cb7882fb1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.532820] env[62000]: DEBUG nova.objects.instance [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'flavor' on Instance uuid d2f87d82-af5e-4b17-959e-40865a8852b3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 940.737289] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 940.765168] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 940.765614] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 940.765876] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 940.766156] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 940.766840] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 940.766840] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 940.767068] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 940.767316] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 940.767622] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 940.767895] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 940.769092] env[62000]: DEBUG nova.virt.hardware [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 940.772849] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a8a8c0-aacd-46cb-a644-e3940db86632 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.786198] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6152201f-2665-41b7-8dcf-c2a523aa8a36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.833556] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.855724] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.878275] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 940.878760] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b188138d-1b6f-4350-81ae-3a6edc6a64b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.892159] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 940.892159] env[62000]: value = "task-882627" [ 940.892159] env[62000]: _type = "Task" [ 940.892159] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.907090] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882627, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.993757] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "refresh_cache-66edb906-f936-4f57-833d-224f36af109e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.014995] env[62000]: INFO nova.compute.manager [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Took 24.83 seconds to build instance. [ 941.040040] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8c403d58-7be9-459d-8320-19725a935ad7 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.339s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.324501] env[62000]: DEBUG nova.compute.manager [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Received event network-vif-plugged-86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 941.324672] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] Acquiring lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.325033] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.325872] env[62000]: DEBUG oslo_concurrency.lockutils [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.326980] env[62000]: DEBUG nova.compute.manager [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] No waiting events found dispatching network-vif-plugged-86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 941.327281] env[62000]: WARNING nova.compute.manager [req-dc990848-a477-4c3d-b4b7-8a255d6c796a req-f6192dd7-87c3-4152-9d73-47b2a6fe6bcd service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Received unexpected event network-vif-plugged-86034d5a-0fb2-425b-a306-aa065f14015a for instance with vm_state building and task_state spawning. [ 941.342163] env[62000]: DEBUG oslo_vmware.api [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882612, 'name': ReconfigVM_Task, 'duration_secs': 5.844505} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.342906] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.343358] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Reconfigured VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 941.361280] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.408921] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882627, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.447188] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Successfully updated port: 86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 941.497698] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 941.498653] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d5e43ab-dffc-4cdd-b15c-6ab1a0af2c35 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.513373] env[62000]: DEBUG oslo_vmware.api [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 941.513373] env[62000]: value = "task-882628" [ 941.513373] env[62000]: _type = "Task" [ 941.513373] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.520666] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f1bd6e00-c07b-44fc-b3d2-35ee77930f29 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.340s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.529130] env[62000]: DEBUG oslo_vmware.api [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882628, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.578434] env[62000]: INFO nova.compute.manager [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Rebuilding instance [ 941.648027] env[62000]: DEBUG nova.compute.manager [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 941.648027] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bed8399-17df-4928-8274-fb0fe9fa13be {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.856469] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.906645] env[62000]: DEBUG oslo_vmware.api [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882627, 'name': PowerOnVM_Task, 'duration_secs': 0.566678} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.907095] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 941.907316] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2e3620ef-b188-4353-9896-5b8fe3e8ad70 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance 'dfa26fab-40ff-4409-bb73-79aa555ab225' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 941.950920] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.951144] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.951353] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 942.014264] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "54eb64ad-bc7b-4495-9e93-451059139c4c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.014617] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.014850] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.015048] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.015231] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.021750] env[62000]: INFO nova.compute.manager [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Terminating instance [ 942.024686] env[62000]: DEBUG nova.compute.manager [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 942.025055] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 942.026328] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7dc5f254-334f-4281-90bf-a290be165ad7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.033305] env[62000]: DEBUG oslo_vmware.api [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882628, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.039286] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 942.039657] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c81263c5-51ed-404b-a0a4-0e504fb76ddc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.048621] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 942.048621] env[62000]: value = "task-882629" [ 942.048621] env[62000]: _type = "Task" [ 942.048621] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.058684] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882629, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.162906] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 942.163369] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-297160a0-af8a-4b42-8d39-80063480f778 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.173925] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 942.173925] env[62000]: value = "task-882630" [ 942.173925] env[62000]: _type = "Task" [ 942.173925] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.184631] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.355726] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.491115] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 942.529676] env[62000]: DEBUG oslo_vmware.api [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882628, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.559017] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882629, 'name': PowerOffVM_Task, 'duration_secs': 0.330967} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.559429] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.559622] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 942.559893] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-34c90d63-02b8-4bd8-b80d-d502ecf2bf65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.654161] env[62000]: DEBUG nova.network.neutron [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Updating instance_info_cache with network_info: [{"id": "86034d5a-0fb2-425b-a306-aa065f14015a", "address": "fa:16:3e:f7:88:88", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86034d5a-0f", "ovs_interfaceid": "86034d5a-0fb2-425b-a306-aa065f14015a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.684828] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882630, 'name': PowerOffVM_Task, 'duration_secs': 0.311359} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.685141] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 942.750549] env[62000]: INFO nova.compute.manager [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Detaching volume 42495829-8539-4f47-9949-eb04eac3dc16 [ 942.787689] env[62000]: INFO nova.virt.block_device [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Attempting to driver detach volume 42495829-8539-4f47-9949-eb04eac3dc16 from mountpoint /dev/sdb [ 942.787968] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 942.792026] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 942.792026] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89c7470-e441-4179-a5a8-7f0788da2db0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.813793] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df38a07-8c9c-417b-a184-23e3f72f89d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.822087] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-431c82dd-2018-4479-ba81-391aefec803e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.843303] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd7b359-30e0-4271-92e1-7bd28f142388 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.864351] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] The volume has not been displaced from its original location: [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 942.869758] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 942.873256] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ff8a399-90a1-4ab0-bb79-df3670633e49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.886495] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882626, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.651696} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.886773] env[62000]: INFO nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0/OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0.vmdk to [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk. [ 942.886969] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Cleaning up location [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 942.887146] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_97b457ce-80fc-4ee2-9436-0beadef861a0 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 942.887865] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b561d921-0d98-44b7-9f19-8f03fb9ed42e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.894409] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 942.894409] env[62000]: value = "task-882632" [ 942.894409] env[62000]: _type = "Task" [ 942.894409] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.898562] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 942.898562] env[62000]: value = "task-882633" [ 942.898562] env[62000]: _type = "Task" [ 942.898562] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.905028] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882632, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.910345] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882633, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.965452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 942.965452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 942.965648] env[62000]: DEBUG nova.network.neutron [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 943.027430] env[62000]: DEBUG oslo_vmware.api [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882628, 'name': PowerOnVM_Task, 'duration_secs': 1.11302} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.027700] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 943.027889] env[62000]: DEBUG nova.compute.manager [None req-1a80e4ea-cce5-4b50-a11a-08583ae971d9 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 943.028674] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f14f9dbe-a94d-4293-9583-aad1e0cd0b01 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.156645] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.156996] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Instance network_info: |[{"id": "86034d5a-0fb2-425b-a306-aa065f14015a", "address": "fa:16:3e:f7:88:88", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86034d5a-0f", "ovs_interfaceid": "86034d5a-0fb2-425b-a306-aa065f14015a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 943.157457] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f7:88:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86034d5a-0fb2-425b-a306-aa065f14015a', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 943.164929] env[62000]: DEBUG oslo.service.loopingcall [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 943.165177] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 943.165436] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c9c86ad9-da1e-4c4d-8011-e638dc78dd14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.186667] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 943.186667] env[62000]: value = "task-882634" [ 943.186667] env[62000]: _type = "Task" [ 943.186667] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.196009] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882634, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.362211] env[62000]: DEBUG nova.compute.manager [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Received event network-changed-86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 943.362506] env[62000]: DEBUG nova.compute.manager [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Refreshing instance network info cache due to event network-changed-86034d5a-0fb2-425b-a306-aa065f14015a. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 943.362771] env[62000]: DEBUG oslo_concurrency.lockutils [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] Acquiring lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 943.362937] env[62000]: DEBUG oslo_concurrency.lockutils [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] Acquired lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 943.363300] env[62000]: DEBUG nova.network.neutron [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Refreshing network info cache for port 86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 943.412937] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882632, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.416186] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882633, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.11593} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.416655] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 943.416990] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 943.417426] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk to [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 943.417826] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68331109-5731-416d-8e92-fe91e8521237 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.438482] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 943.438482] env[62000]: value = "task-882635" [ 943.438482] env[62000]: _type = "Task" [ 943.438482] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.450154] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.699432] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882634, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.893247] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 943.893537] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 943.893729] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore1] 54eb64ad-bc7b-4495-9e93-451059139c4c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 943.894070] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-58228b77-54b0-4def-b606-4c5e82883a4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.908214] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 943.908214] env[62000]: value = "task-882636" [ 943.908214] env[62000]: _type = "Task" [ 943.908214] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.912318] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882632, 'name': ReconfigVM_Task, 'duration_secs': 0.544546} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.916075] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.921095] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9d5be75c-8d7a-4316-9081-7b5b25b1ea56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.938790] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882636, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.943764] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 943.943764] env[62000]: value = "task-882637" [ 943.943764] env[62000]: _type = "Task" [ 943.943764] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.951402] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.958852] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882637, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.204265] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882634, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.422118] env[62000]: DEBUG oslo_vmware.api [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882636, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266034} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.422445] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 944.422643] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 944.422824] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 944.423015] env[62000]: INFO nova.compute.manager [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Took 2.40 seconds to destroy the instance on the hypervisor. [ 944.423278] env[62000]: DEBUG oslo.service.loopingcall [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 944.423484] env[62000]: DEBUG nova.compute.manager [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 944.423582] env[62000]: DEBUG nova.network.neutron [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 944.451081] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.458083] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882637, 'name': ReconfigVM_Task, 'duration_secs': 0.236182} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.458382] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 944.592027] env[62000]: INFO nova.network.neutron [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Port 68fd0d34-536e-4504-a837-62eca9533e5f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 944.592909] env[62000]: DEBUG nova.network.neutron [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.701469] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882634, 'name': CreateVM_Task, 'duration_secs': 1.500442} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.703949] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 944.704530] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.704780] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.705105] env[62000]: DEBUG nova.compute.manager [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Going to confirm migration 4 {{(pid=62000) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 944.710062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.710062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.710062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 944.710062] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20bd1260-c038-40f8-a9e0-5e7860523e48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.715768] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 944.715768] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ca146-ffde-5f3c-c909-59f6a0baec47" [ 944.715768] env[62000]: _type = "Task" [ 944.715768] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.725946] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ca146-ffde-5f3c-c909-59f6a0baec47, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.733467] env[62000]: DEBUG nova.network.neutron [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Updated VIF entry in instance network info cache for port 86034d5a-0fb2-425b-a306-aa065f14015a. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 944.734193] env[62000]: DEBUG nova.network.neutron [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Updating instance_info_cache with network_info: [{"id": "86034d5a-0fb2-425b-a306-aa065f14015a", "address": "fa:16:3e:f7:88:88", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86034d5a-0f", "ovs_interfaceid": "86034d5a-0fb2-425b-a306-aa065f14015a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 944.950368] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.096303] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.235087] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ca146-ffde-5f3c-c909-59f6a0baec47, 'name': SearchDatastore_Task, 'duration_secs': 0.083223} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.235087] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.235087] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 945.235087] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.235087] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.235087] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 945.235087] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-397a019c-57d6-4df6-9192-c4aba3efee6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.237912] env[62000]: DEBUG oslo_concurrency.lockutils [req-791ce25a-c811-4138-93f2-3989a57e081a req-3f1e095b-4000-4209-89d2-80594eac2b20 service nova] Releasing lock "refresh_cache-ea58da8c-e4ac-4863-942d-6294cf04e6f2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.257021] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 945.257021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 945.257021] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3705455-d1c0-4b83-bf6d-ba2322c7fe91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.265632] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 945.265632] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52454bfb-14ae-666d-c46b-da9dcd356924" [ 945.265632] env[62000]: _type = "Task" [ 945.265632] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.270842] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.270842] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.270842] env[62000]: DEBUG nova.network.neutron [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 945.270842] env[62000]: DEBUG nova.objects.instance [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'info_cache' on Instance uuid dfa26fab-40ff-4409-bb73-79aa555ab225 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.282242] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52454bfb-14ae-666d-c46b-da9dcd356924, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.449160] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.496791] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 945.497037] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing instance network info cache due to event network-changed-f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 945.497266] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Acquiring lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.497416] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Acquired lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.497581] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Refreshing network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 945.522437] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 945.523140] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a01b1323-5138-426f-916c-50e30f88c636 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.533340] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 945.533340] env[62000]: value = "task-882638" [ 945.533340] env[62000]: _type = "Task" [ 945.533340] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.542671] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882638, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.600481] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cc7a53e-8180-4cb9-958a-5d6a3325cebc tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-72a5bf70-dc6e-4887-abb8-8fbad64bb065-68fd0d34-536e-4504-a837-62eca9533e5f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.664s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.625972] env[62000]: DEBUG nova.network.neutron [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.783364] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52454bfb-14ae-666d-c46b-da9dcd356924, 'name': SearchDatastore_Task, 'duration_secs': 0.076881} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.785054] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef979676-df51-45cf-b485-6fc647676547 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.790166] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.790480] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.790928] env[62000]: DEBUG nova.objects.instance [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'flavor' on Instance uuid 94605ab8-8167-430e-b1cd-c8f51e50d8b9 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 945.797738] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 945.797738] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52db436b-25b5-8ef0-cd1a-82b7f2c74238" [ 945.797738] env[62000]: _type = "Task" [ 945.797738] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.807551] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52db436b-25b5-8ef0-cd1a-82b7f2c74238, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.950590] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.045845] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 946.045845] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 946.045845] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 946.046612] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3a9464-510a-4ced-ae1b-b8a6e8dd332e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.066858] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22821476-6e43-4491-a88e-fdc0dc3530e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.074394] env[62000]: WARNING nova.virt.vmwareapi.driver [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] The volume None does not exist!: nova.exception.DiskNotFound: Unable to find volume [ 946.074734] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 946.077514] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cecd2a-52ff-41d5-9d9f-1ac19bfdbb0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.084862] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 946.084862] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ba676b9c-a420-4f42-a4de-4ef487c7cb14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.129069] env[62000]: INFO nova.compute.manager [-] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Took 1.71 seconds to deallocate network for instance. [ 946.168532] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 946.168747] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 946.169333] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.169333] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d560b32-4529-4ee8-a071-9736f2945b9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.178898] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 946.178898] env[62000]: value = "task-882640" [ 946.178898] env[62000]: _type = "Task" [ 946.178898] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.189158] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.280075] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updated VIF entry in instance network info cache for port f0d1137e-2373-4645-85db-8c5c8c963ff4. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 946.280382] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [{"id": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "address": "fa:16:3e:7d:b4:87", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf0d1137e-23", "ovs_interfaceid": "f0d1137e-2373-4645-85db-8c5c8c963ff4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.310266] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52db436b-25b5-8ef0-cd1a-82b7f2c74238, 'name': SearchDatastore_Task, 'duration_secs': 0.078646} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.310542] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.311105] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ea58da8c-e4ac-4863-942d-6294cf04e6f2/ea58da8c-e4ac-4863-942d-6294cf04e6f2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 946.311323] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8363955a-87d4-4f8b-94e5-3b0afdf8361a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.319506] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 946.319506] env[62000]: value = "task-882641" [ 946.319506] env[62000]: _type = "Task" [ 946.319506] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.329531] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.451554] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882635, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.607477} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.451833] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/fec7d673-ed56-4ee2-8361-5001b5b340d5/fec7d673-ed56-4ee2-8361-5001b5b340d5.vmdk to [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 946.452800] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b55e430-d0fc-426f-9f81-7d14a669a437 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.475694] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 946.475999] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a8d67959-0105-4544-8e34-04c4e49e5bda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.499445] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 946.499445] env[62000]: value = "task-882642" [ 946.499445] env[62000]: _type = "Task" [ 946.499445] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.512604] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882642, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.527710] env[62000]: DEBUG nova.objects.instance [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'pci_requests' on Instance uuid 94605ab8-8167-430e-b1cd-c8f51e50d8b9 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.635892] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.636246] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.636489] env[62000]: DEBUG nova.objects.instance [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'resources' on Instance uuid 54eb64ad-bc7b-4495-9e93-451059139c4c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 946.661863] env[62000]: DEBUG nova.network.neutron [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [{"id": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "address": "fa:16:3e:41:7c:20", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap65be2bd7-e8", "ovs_interfaceid": "65be2bd7-e84a-4cc8-b937-6140dc13b2e4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.690732] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882640, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.784858] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Releasing lock "refresh_cache-72a5bf70-dc6e-4887-abb8-8fbad64bb065" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.785228] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 946.785324] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing instance network info cache due to event network-changed-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 946.785541] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.785687] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.785854] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 946.829195] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.010931] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.030767] env[62000]: DEBUG nova.objects.base [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Object Instance<94605ab8-8167-430e-b1cd-c8f51e50d8b9> lazy-loaded attributes: flavor,pci_requests {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 947.031111] env[62000]: DEBUG nova.network.neutron [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 947.093534] env[62000]: DEBUG nova.policy [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a908ddc4430b471a8023f73ffed5df0b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '747546b09fd04a41b9c2df860a699186', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 947.164302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-dfa26fab-40ff-4409-bb73-79aa555ab225" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.164727] env[62000]: DEBUG nova.objects.instance [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'migration_context' on Instance uuid dfa26fab-40ff-4409-bb73-79aa555ab225 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.194513] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882640, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.746481} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.194815] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 947.195144] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 947.195237] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 947.335892] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882641, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.404501] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-741ffb6b-c2a4-4f7f-8fdd-c70811e947a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.413385] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16b4c2df-4422-431f-9e85-aa8120e4c5dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.444374] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73f73a88-c5d7-4c54-91e6-35258c508c23 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.455256] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cc27d35-7e8b-429b-8012-6e773aba46e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.471651] env[62000]: DEBUG nova.compute.provider_tree [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 947.511797] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882642, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.670233] env[62000]: DEBUG nova.objects.base [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 947.671217] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3722b0ad-db21-4754-a3c2-2c2f61967727 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.677300] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updated VIF entry in instance network info cache for port 1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 947.677648] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.694202] env[62000]: DEBUG oslo_concurrency.lockutils [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.694347] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Received event network-vif-deleted-5b734604-c3c4-4667-851c-a635f7aeafc6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 947.695552] env[62000]: INFO nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Neutron deleted interface 5b734604-c3c4-4667-851c-a635f7aeafc6; detaching it from the instance and deleting it from the info cache [ 947.695552] env[62000]: DEBUG nova.network.neutron [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.695742] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e5dab0d-f86c-4405-b5de-8a5306f1f26f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.702811] env[62000]: INFO nova.virt.block_device [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Booting with volume 42495829-8539-4f47-9949-eb04eac3dc16 at /dev/sdb [ 947.706949] env[62000]: DEBUG oslo_vmware.api [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 947.706949] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525941e2-5617-f841-0053-0c6115ad841c" [ 947.706949] env[62000]: _type = "Task" [ 947.706949] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.719132] env[62000]: DEBUG oslo_vmware.api [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]525941e2-5617-f841-0053-0c6115ad841c, 'name': SearchDatastore_Task, 'duration_secs': 0.010393} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.719463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.747586] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b463084-226f-44bf-a917-b4089210c9ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.758807] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926bf795-c26a-44c0-ba87-79998f379241 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.806031] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ecb3faac-5473-49f0-8cc1-8caf310d61b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.815848] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7538cf1d-3c8b-4045-a1ef-f20e1e96d2e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.837027] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882641, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.107521} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.837252] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] ea58da8c-e4ac-4863-942d-6294cf04e6f2/ea58da8c-e4ac-4863-942d-6294cf04e6f2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 947.837479] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 947.837733] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c4579bfd-ebee-4bae-97a3-09e93221dcc6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.845587] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 947.845587] env[62000]: value = "task-882643" [ 947.845587] env[62000]: _type = "Task" [ 947.845587] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.861852] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2fe933-5ea2-40e1-99dc-30b550b36784 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.871572] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882643, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.872324] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff8a024-f907-4035-a5a6-93f0a39785c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.886824] env[62000]: DEBUG nova.virt.block_device [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating existing volume attachment record: 78f6bc6a-204b-4e3a-a95f-99040e9bbf81 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 947.974890] env[62000]: DEBUG nova.scheduler.client.report [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 948.012013] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882642, 'name': ReconfigVM_Task, 'duration_secs': 1.03039} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.012406] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400/8a11689f-fc00-43f8-9215-8d81daa84400.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.014008] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encrypted': False, 'guest_format': None, 'encryption_format': None, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'boot_index': 0, 'size': 0, 'image_id': '37b135ac-1e3f-46c1-bb0e-ce7f0b69db72'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '450fe1c5-d3c7-40cf-b740-8b954b02dec8', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201608', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'name': 'volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8a11689f-fc00-43f8-9215-8d81daa84400', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'serial': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5'}, 'mount_device': '/dev/sdb', 'device_type': None, 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62000) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 948.014186] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 948.014510] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201608', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'name': 'volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8a11689f-fc00-43f8-9215-8d81daa84400', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'serial': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 948.015371] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05c01315-0e87-40b4-b317-bd6fa54380c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.033498] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd4ff3bf-4c77-409e-8f4a-430d10b0572f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.058892] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5/volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.059281] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6202d7e9-c672-48c3-9108-45b3c8f097f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.079874] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 948.079874] env[62000]: value = "task-882644" [ 948.079874] env[62000]: _type = "Task" [ 948.079874] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.089311] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882644, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.199898] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e0433c5a-9c49-4e90-8a89-513e433bb9a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.210465] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17fe9ac0-7122-4b5b-a491-18571e928823 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.246335] env[62000]: DEBUG nova.compute.manager [req-98b064e4-8b2c-4aa2-864c-2dc590b7a77e req-7a6024db-fe48-436f-ad12-d6da502cec38 service nova] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Detach interface failed, port_id=5b734604-c3c4-4667-851c-a635f7aeafc6, reason: Instance 54eb64ad-bc7b-4495-9e93-451059139c4c could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 948.368576] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882643, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.112297} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.368807] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 948.369604] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68d526eb-e169-4938-b6f5-9d7d492a3b99 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.392311] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] ea58da8c-e4ac-4863-942d-6294cf04e6f2/ea58da8c-e4ac-4863-942d-6294cf04e6f2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 948.392649] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c350f3aa-c824-4613-a330-a0d6b44fdcf2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.413289] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 948.413289] env[62000]: value = "task-882645" [ 948.413289] env[62000]: _type = "Task" [ 948.413289] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.421187] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882645, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.479798] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 948.482725] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.763s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 948.511780] env[62000]: INFO nova.scheduler.client.report [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted allocations for instance 54eb64ad-bc7b-4495-9e93-451059139c4c [ 948.592222] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.924054] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882645, 'name': ReconfigVM_Task, 'duration_secs': 0.286228} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.924378] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Reconfigured VM instance instance-00000059 to attach disk [datastore2] ea58da8c-e4ac-4863-942d-6294cf04e6f2/ea58da8c-e4ac-4863-942d-6294cf04e6f2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 948.925093] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ac372c00-e6b3-4ad1-8dd5-85d49ffda0c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.934696] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 948.934696] env[62000]: value = "task-882646" [ 948.934696] env[62000]: _type = "Task" [ 948.934696] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.944428] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882646, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.023195] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2b8dbe63-1080-4cb9-bcf6-c849b9e3c45c tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "54eb64ad-bc7b-4495-9e93-451059139c4c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.008s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.093994] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882644, 'name': ReconfigVM_Task, 'duration_secs': 0.87472} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.094455] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfigured VM instance instance-00000044 to attach disk [datastore1] volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5/volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 949.110713] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-eeb5120a-038a-49dc-b1f6-2c3cac239133 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.128712] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 949.128712] env[62000]: value = "task-882647" [ 949.128712] env[62000]: _type = "Task" [ 949.128712] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.138600] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882647, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.243582] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbcf2e6f-8bfc-429f-8262-9333530f91f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.253581] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eece85ae-af07-4837-af10-6f093129e512 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.286638] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1feacc0b-49ed-4b77-9ebf-26004ee2b747 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.295143] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-175d061b-322e-4d77-aa32-98c5307361e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.309258] env[62000]: DEBUG nova.compute.provider_tree [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.420645] env[62000]: DEBUG nova.network.neutron [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Successfully updated port: 68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 949.445991] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882646, 'name': Rename_Task, 'duration_secs': 0.156982} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.446484] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 949.446765] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-84fb3784-b697-4401-99ff-1d16137204c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.454949] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 949.454949] env[62000]: value = "task-882648" [ 949.454949] env[62000]: _type = "Task" [ 949.454949] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.463812] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882648, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.491470] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "c354ba41-0915-44cc-9857-42705a56c00f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.491864] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.492186] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "c354ba41-0915-44cc-9857-42705a56c00f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.492402] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.492614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.495093] env[62000]: INFO nova.compute.manager [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Terminating instance [ 949.497390] env[62000]: DEBUG nova.compute.manager [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 949.497663] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 949.498599] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-677f9779-9b60-4145-b662-7b778f0df9c6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.509798] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 949.510063] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e82b891a-198f-4d01-aa0c-debb96d0f54e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.518673] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 949.518673] env[62000]: value = "task-882649" [ 949.518673] env[62000]: _type = "Task" [ 949.518673] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.528322] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882649, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.642329] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882647, 'name': ReconfigVM_Task, 'duration_secs': 0.229782} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.642738] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201608', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'name': 'volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8a11689f-fc00-43f8-9215-8d81daa84400', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'serial': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 949.643474] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-389704b1-db7c-4d54-9784-02a776901258 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.652798] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 949.652798] env[62000]: value = "task-882650" [ 949.652798] env[62000]: _type = "Task" [ 949.652798] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.662914] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882650, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.789192] env[62000]: DEBUG nova.compute.manager [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 949.789336] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.789882] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.789882] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.790126] env[62000]: DEBUG nova.compute.manager [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] No waiting events found dispatching network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 949.790126] env[62000]: WARNING nova.compute.manager [req-7f2cc81e-a99b-4d54-ad58-7745003851fe req-00032749-7e67-4bab-98ef-10691f659ecc service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received unexpected event network-vif-plugged-68fd0d34-536e-4504-a837-62eca9533e5f for instance with vm_state active and task_state None. [ 949.812480] env[62000]: DEBUG nova.scheduler.client.report [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.923275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 949.923432] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 949.923644] env[62000]: DEBUG nova.network.neutron [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 949.966859] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882648, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.004272] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 950.004602] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 950.004807] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 950.005111] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 950.005215] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 950.005373] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 950.005589] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 950.005751] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 950.005921] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 950.006104] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 950.006324] env[62000]: DEBUG nova.virt.hardware [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 950.007261] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f2f0af-595c-415c-a496-3482226fda1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.016132] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6089e5c1-b2b8-4501-a996-3fe6e0abb3d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.037521] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:86:36:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '919b3f5d-af5d-45e1-896f-ac0ea1309017', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 950.045229] env[62000]: DEBUG oslo.service.loopingcall [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.045569] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882649, 'name': PowerOffVM_Task, 'duration_secs': 0.260003} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.045809] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 950.046105] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 950.046300] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 950.046566] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3bb01ab8-45a2-4287-adfb-163e46d3dd59 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.060882] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-81be3530-2ed9-4ea7-a0dd-2de342205f6f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.070365] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 950.070365] env[62000]: value = "task-882652" [ 950.070365] env[62000]: _type = "Task" [ 950.070365] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.078970] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882652, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.166294] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882650, 'name': Rename_Task, 'duration_secs': 0.225485} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.167723] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 950.168081] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 950.168265] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 950.168466] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore1] c354ba41-0915-44cc-9857-42705a56c00f {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 950.168723] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-13073078-884b-4f71-bfd6-b51771121bec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.170592] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff0fb88e-f696-4604-880a-7960e7c8ff70 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.179764] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 950.179764] env[62000]: value = "task-882653" [ 950.179764] env[62000]: _type = "Task" [ 950.179764] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.181315] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 950.181315] env[62000]: value = "task-882654" [ 950.181315] env[62000]: _type = "Task" [ 950.181315] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.197736] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882654, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.201889] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882653, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.269718] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.269999] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.466932] env[62000]: DEBUG oslo_vmware.api [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882648, 'name': PowerOnVM_Task, 'duration_secs': 0.553889} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.467312] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 950.467572] env[62000]: INFO nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Took 9.73 seconds to spawn the instance on the hypervisor. [ 950.467790] env[62000]: DEBUG nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 950.468660] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57293226-da1b-4d7a-865c-1d83c358d500 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.490372] env[62000]: WARNING nova.network.neutron [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] e21c3795-1c6b-42ef-af81-e113912fa80d already exists in list: networks containing: ['e21c3795-1c6b-42ef-af81-e113912fa80d']. ignoring it [ 950.581768] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882652, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.694547] env[62000]: DEBUG oslo_vmware.api [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882654, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.312487} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.697530] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.697753] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 950.697941] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 950.698184] env[62000]: INFO nova.compute.manager [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Took 1.20 seconds to destroy the instance on the hypervisor. [ 950.698466] env[62000]: DEBUG oslo.service.loopingcall [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.698679] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882653, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.701384] env[62000]: DEBUG nova.compute.manager [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.701514] env[62000]: DEBUG nova.network.neutron [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 950.772616] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 950.823244] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.340s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.989244] env[62000]: INFO nova.compute.manager [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Took 14.70 seconds to build instance. [ 951.005363] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 951.005670] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 951.005746] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 951.082051] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882652, 'name': CreateVM_Task, 'duration_secs': 0.668306} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.082319] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 951.083043] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.083268] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.083648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 951.083926] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e80665f4-769b-4a4c-9e96-35dbc0eb02fc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.090181] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 951.090181] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5201b18f-2a4d-84ce-487c-5cb289567607" [ 951.090181] env[62000]: _type = "Task" [ 951.090181] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.099505] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5201b18f-2a4d-84ce-487c-5cb289567607, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.195026] env[62000]: DEBUG oslo_vmware.api [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882653, 'name': PowerOnVM_Task, 'duration_secs': 0.980716} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.195430] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 951.248910] env[62000]: DEBUG nova.network.neutron [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "68fd0d34-536e-4504-a837-62eca9533e5f", "address": "fa:16:3e:01:88:92", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fd0d34-53", "ovs_interfaceid": "68fd0d34-536e-4504-a837-62eca9533e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.296932] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.297262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.298963] env[62000]: INFO nova.compute.claims [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 951.328111] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 951.331487] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9024bbdb-16c4-4114-8c07-cde7b5b31f82 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.335045] env[62000]: DEBUG nova.compute.manager [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 951.335655] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96015856-100e-402b-a8f0-ef882341de05 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.355316] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 951.355316] env[62000]: ERROR oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk due to incomplete transfer. [ 951.356468] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-84358a90-22c7-4cc6-8bb8-ad2dc75d49a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.365854] env[62000]: DEBUG oslo_vmware.rw_handles [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/529ab607-e684-828c-6b6c-d3f97f72864d/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 951.366419] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Uploaded image 73acd51d-de92-4df9-a9a6-9faafd098410 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 951.369031] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 951.369620] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-754d4932-ce06-43c5-800f-17299f58503d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.377539] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 951.377539] env[62000]: value = "task-882655" [ 951.377539] env[62000]: _type = "Task" [ 951.377539] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.387072] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882655, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.415283] env[62000]: INFO nova.scheduler.client.report [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocation for migration 3ce7c346-1ce0-403a-9471-e10958a315c3 [ 951.492115] env[62000]: DEBUG oslo_concurrency.lockutils [None req-011fafdc-1198-4aa1-8f1d-17277f396bbd tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.220s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.602379] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5201b18f-2a4d-84ce-487c-5cb289567607, 'name': SearchDatastore_Task, 'duration_secs': 0.013186} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.602732] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.603015] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 951.603307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.604030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.604030] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 951.604030] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ab156c8a-0dc2-4793-81e1-ba604c327a72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.615672] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 951.615864] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 951.616614] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0b20832-dd8b-4d6c-b2e3-e8a293736105 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.622153] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 951.622153] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc8dd-0748-5b65-72ad-307c66279465" [ 951.622153] env[62000]: _type = "Task" [ 951.622153] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.630028] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc8dd-0748-5b65-72ad-307c66279465, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.752601] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.753288] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.753458] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.754328] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d2300bd-034e-4134-be48-3ef7d10b19df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.771362] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.771662] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.772493] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.772795] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.773019] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.775244] env[62000]: DEBUG nova.virt.hardware [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.781588] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfiguring VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1929}} [ 951.782377] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67dc0daa-af60-48e2-b539-0051c18590e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.801736] env[62000]: DEBUG oslo_vmware.api [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 951.801736] env[62000]: value = "task-882656" [ 951.801736] env[62000]: _type = "Task" [ 951.801736] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.813310] env[62000]: DEBUG oslo_vmware.api [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882656, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.863875] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06923330-9a61-4689-97de-e2a8acf7ee14 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 35.737s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.889784] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882655, 'name': Destroy_Task, 'duration_secs': 0.370913} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.890685] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Destroyed the VM [ 951.890942] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 951.891222] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-38721fee-5f5c-4a1e-b88c-9088530ccba5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.895646] env[62000]: DEBUG nova.compute.manager [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-changed-68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 951.896220] env[62000]: DEBUG nova.compute.manager [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing instance network info cache due to event network-changed-68fd0d34-536e-4504-a837-62eca9533e5f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 951.896220] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.896220] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.898292] env[62000]: DEBUG nova.network.neutron [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Refreshing network info cache for port 68fd0d34-536e-4504-a837-62eca9533e5f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 951.906225] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 951.906225] env[62000]: value = "task-882657" [ 951.906225] env[62000]: _type = "Task" [ 951.906225] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.916871] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882657, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.920260] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8ff7a793-2a9a-4a13-8b2e-85df1ee88a9e tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.215s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.990223] env[62000]: DEBUG nova.network.neutron [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.136776] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526fc8dd-0748-5b65-72ad-307c66279465, 'name': SearchDatastore_Task, 'duration_secs': 0.027294} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.137682] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07c2868e-262c-4e11-92d1-cacf6359827d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.145828] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 952.145828] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522de62c-2ae0-896b-74f0-2ca44308d33a" [ 952.145828] env[62000]: _type = "Task" [ 952.145828] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.156563] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522de62c-2ae0-896b-74f0-2ca44308d33a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.315103] env[62000]: DEBUG oslo_vmware.api [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882656, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.417160] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882657, 'name': RemoveSnapshot_Task, 'duration_secs': 0.501024} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.419373] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 952.419672] env[62000]: DEBUG nova.compute.manager [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 952.421105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ce78b8-3254-48e4-807c-ce1ab6eeb1d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.494571] env[62000]: INFO nova.compute.manager [-] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Took 1.79 seconds to deallocate network for instance. [ 952.540976] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50200928-b4a7-49ea-a214-38fae3ffa8c5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.549797] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb2aa069-c233-4f59-9459-1f15d58db706 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.586277] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb59fd2-d9ad-4444-951a-ceb94731f961 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.596027] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749c443e-24af-45aa-9638-1b24037e88cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.609631] env[62000]: DEBUG nova.compute.provider_tree [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 952.658309] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522de62c-2ae0-896b-74f0-2ca44308d33a, 'name': SearchDatastore_Task, 'duration_secs': 0.016071} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.658449] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.658726] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 952.658992] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c1d17681-8ac9-41f0-b13e-de2e2b119fdd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.669126] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 952.669126] env[62000]: value = "task-882658" [ 952.669126] env[62000]: _type = "Task" [ 952.669126] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.678043] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882658, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.790391] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "23366d62-294b-47c8-adc6-80e854f452ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.790705] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 952.792842] env[62000]: DEBUG nova.network.neutron [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updated VIF entry in instance network info cache for port 68fd0d34-536e-4504-a837-62eca9533e5f. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 952.793337] env[62000]: DEBUG nova.network.neutron [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "68fd0d34-536e-4504-a837-62eca9533e5f", "address": "fa:16:3e:01:88:92", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap68fd0d34-53", "ovs_interfaceid": "68fd0d34-536e-4504-a837-62eca9533e5f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.814256] env[62000]: DEBUG oslo_vmware.api [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882656, 'name': ReconfigVM_Task, 'duration_secs': 0.829737} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.814920] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.815241] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfigured VM to attach interface {{(pid=62000) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1943}} [ 952.934787] env[62000]: INFO nova.compute.manager [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Shelve offloading [ 952.936746] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 952.937123] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4fa24662-8310-4848-ab44-a242aa52184c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.945563] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 952.945563] env[62000]: value = "task-882659" [ 952.945563] env[62000]: _type = "Task" [ 952.945563] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.954573] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.000537] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.114690] env[62000]: DEBUG nova.scheduler.client.report [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.182180] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882658, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.295927] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 953.299041] env[62000]: DEBUG oslo_concurrency.lockutils [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.299318] env[62000]: DEBUG nova.compute.manager [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Received event network-vif-deleted-e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 953.299505] env[62000]: INFO nova.compute.manager [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Neutron deleted interface e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425; detaching it from the instance and deleting it from the info cache [ 953.299683] env[62000]: DEBUG nova.network.neutron [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.320290] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e2122671-5129-46c9-96ff-3e208466842b tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.530s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.456136] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 953.456359] env[62000]: DEBUG nova.compute.manager [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.457267] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04159a45-9636-4e4c-a5e6-07d90c9f5ea5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.464497] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.464669] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.464853] env[62000]: DEBUG nova.network.neutron [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 953.581367] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.581654] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.581889] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.582118] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.582307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.584639] env[62000]: INFO nova.compute.manager [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Terminating instance [ 953.586534] env[62000]: DEBUG nova.compute.manager [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 953.586734] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 953.587574] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3549cd0a-cee9-44d3-b5f7-6ecc33013ce3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.596230] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 953.596476] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ac0e5878-0546-4ccc-b854-74b6bef506ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.603702] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 953.603702] env[62000]: value = "task-882660" [ 953.603702] env[62000]: _type = "Task" [ 953.603702] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.613190] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882660, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.620218] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.323s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 953.620774] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 953.623449] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.623s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.623679] env[62000]: DEBUG nova.objects.instance [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'resources' on Instance uuid c354ba41-0915-44cc-9857-42705a56c00f {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.680982] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882658, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.806845] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8ed33dc4-9625-4e7e-bc03-f77fb4567f86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.817776] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-647662bc-572d-4fa4-b9f4-928888478e18 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.830473] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.857118] env[62000]: DEBUG nova.compute.manager [req-a90aeda1-c228-4a43-b892-794cf28e0649 req-e312124f-1fcc-4b65-af47-0b3694602639 service nova] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Detach interface failed, port_id=e48f9f01-2f64-4bad-bfc4-3d5fdf4cc425, reason: Instance c354ba41-0915-44cc-9857-42705a56c00f could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 954.083321] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.083664] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.114948] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882660, 'name': PowerOffVM_Task, 'duration_secs': 0.178126} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.115355] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 954.115603] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 954.115873] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-defc6069-31f4-46eb-ac7f-f88c55443df4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.126906] env[62000]: DEBUG nova.compute.utils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 954.132578] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 954.132981] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 954.185076] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882658, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.207967] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 954.208255] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 954.208445] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore2] dfa26fab-40ff-4409-bb73-79aa555ab225 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 954.208773] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-990ecc90-0274-4b60-a03f-bfe984ee17db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.223142] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 954.223142] env[62000]: value = "task-882662" [ 954.223142] env[62000]: _type = "Task" [ 954.223142] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.230518] env[62000]: DEBUG nova.policy [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ffd8327dd8c4844b776982e99de5e41', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd14aec090f2a4abc8078df1fe4428bbb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 954.237292] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.290971] env[62000]: DEBUG nova.network.neutron [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updating instance_info_cache with network_info: [{"id": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "address": "fa:16:3e:4c:38:15", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape041eaad-7b", "ovs_interfaceid": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 954.365617] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1c3a3b-8021-4d87-9a71-15f8810899de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.374168] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a2628ac-1584-49c9-844a-d321738ec9d9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.406138] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cd96ad-dae0-4b67-9ae6-731132f198b2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.414411] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e156ac41-d55e-4486-8ec6-29f310079184 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.432628] env[62000]: DEBUG nova.compute.provider_tree [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.586831] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 954.600569] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Successfully created port: fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 954.633480] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 954.683149] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882658, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.792919} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.683433] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 954.683739] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 954.684016] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6f7b1c79-948c-4ec3-bbe4-65a0d28bc78a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.692948] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 954.692948] env[62000]: value = "task-882663" [ 954.692948] env[62000]: _type = "Task" [ 954.692948] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.702869] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882663, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.737011] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882662, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.796961] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.936884] env[62000]: DEBUG nova.scheduler.client.report [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.032741] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Didn't find any instances for network info cache update. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 955.032988] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033176] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033330] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033480] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033623] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033770] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.033902] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 955.034061] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 955.061491] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.061756] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.100737] env[62000]: DEBUG nova.compute.manager [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received event network-vif-unplugged-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 955.100987] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.101204] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.101422] env[62000]: DEBUG oslo_concurrency.lockutils [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.101550] env[62000]: DEBUG nova.compute.manager [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] No waiting events found dispatching network-vif-unplugged-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 955.101720] env[62000]: WARNING nova.compute.manager [req-cc2c9517-9abc-4ba5-850a-5fac97702f79 req-d394479d-d47d-4a9e-b9e7-d11541b504fc service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received unexpected event network-vif-unplugged-e041eaad-7b03-44b1-be56-0c3871c66ecc for instance with vm_state shelved and task_state shelving_offloading. [ 955.104856] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.123834] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 955.124800] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0bd158-f33b-4d5a-be67-a2205e2f66fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.135491] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 955.135770] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b83b1d4b-556a-4fc8-95ce-ce398fdcb9fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.203659] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882663, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071864} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.204159] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.205014] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2653ef9-7e95-4723-927e-344112c93d44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.227549] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.229051] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dc08e45-1ff4-4833-8d01-ad269d8f6f91 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.247259] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 955.247485] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 955.247665] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore1] 0cbc29c0-b817-4960-9a8d-5fef0b439d58 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 955.248385] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3ee082cd-0f0a-4ac2-8764-9726ba5d9dcc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.255695] env[62000]: DEBUG oslo_vmware.api [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882662, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.776539} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.256991] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.257173] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.257363] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.257544] env[62000]: INFO nova.compute.manager [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Took 1.67 seconds to destroy the instance on the hypervisor. [ 955.257787] env[62000]: DEBUG oslo.service.loopingcall [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 955.258055] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 955.258055] env[62000]: value = "task-882665" [ 955.258055] env[62000]: _type = "Task" [ 955.258055] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.259117] env[62000]: DEBUG nova.compute.manager [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 955.259239] env[62000]: DEBUG nova.network.neutron [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 955.261177] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 955.261177] env[62000]: value = "task-882666" [ 955.261177] env[62000]: _type = "Task" [ 955.261177] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.273568] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882666, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.276687] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882665, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.442462] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.819s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.445086] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.615s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.446801] env[62000]: INFO nova.compute.claims [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 955.467293] env[62000]: INFO nova.scheduler.client.report [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted allocations for instance c354ba41-0915-44cc-9857-42705a56c00f [ 955.536993] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.564137] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.564414] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.565305] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75797ebb-c3cc-4a9f-864a-fc11ef1fc18f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.583382] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18f3ddc2-6023-4b0d-99fe-6af936263c3a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.610257] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfiguring VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 955.610563] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b422d65f-d7f1-41cc-9e24-bbd58defb3a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.632471] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 955.632471] env[62000]: value = "task-882667" [ 955.632471] env[62000]: _type = "Task" [ 955.632471] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.641585] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.645869] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 955.673559] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 955.673814] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 955.673973] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 955.674234] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 955.674409] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 955.674562] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 955.674765] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 955.674925] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 955.675109] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 955.675378] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 955.675659] env[62000]: DEBUG nova.virt.hardware [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 955.676745] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d18d08-b8b4-44b6-aedd-8d69e8a5e138 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.685326] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-841e85c3-bbf2-458f-8c95-7fdb0c861dfe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.774097] env[62000]: DEBUG oslo_vmware.api [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882666, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165428} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.776909] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 955.777128] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 955.777317] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 955.779184] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882665, 'name': ReconfigVM_Task, 'duration_secs': 0.358367} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.779628] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to attach disk [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3/d2f87d82-af5e-4b17-959e-40865a8852b3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 955.780748] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Block device information present: {'root_device_name': '/dev/sda', 'image': [{'encryption_options': None, 'device_type': 'disk', 'disk_bus': None, 'encrypted': False, 'guest_format': None, 'encryption_format': None, 'device_name': '/dev/sda', 'encryption_secret_uuid': None, 'boot_index': 0, 'size': 0, 'image_id': '37b135ac-1e3f-46c1-bb0e-ce7f0b69db72'}], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': '78f6bc6a-204b-4e3a-a95f-99040e9bbf81', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'}, 'mount_device': '/dev/sdb', 'device_type': None, 'disk_bus': None, 'delete_on_termination': False, 'boot_index': None, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62000) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 955.780964] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 955.781173] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 955.781916] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bfd4f83-dccc-4614-b266-d0512682bc4d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.798246] env[62000]: INFO nova.scheduler.client.report [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance 0cbc29c0-b817-4960-9a8d-5fef0b439d58 [ 955.801131] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3caa1d3-e43b-4b1c-9bc2-4476e22a604e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.826302] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.826616] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a0aeccc-5864-4f9e-8657-198e491c6f69 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.847198] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 955.847198] env[62000]: value = "task-882668" [ 955.847198] env[62000]: _type = "Task" [ 955.847198] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.855944] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882668, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.979052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ef0cf51f-6cec-4ae9-9da4-f2278996e14e tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "c354ba41-0915-44cc-9857-42705a56c00f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.487s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.022943] env[62000]: DEBUG nova.network.neutron [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.146796] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.288124] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Successfully updated port: fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 956.305238] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.356767] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882668, 'name': ReconfigVM_Task, 'duration_secs': 0.374807} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.357083] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to attach disk [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.361685] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b36c7bd5-0470-4398-9f49-2bc443a3b83a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.380575] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 956.380575] env[62000]: value = "task-882669" [ 956.380575] env[62000]: _type = "Task" [ 956.380575] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.389742] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882669, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.525389] env[62000]: INFO nova.compute.manager [-] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Took 1.27 seconds to deallocate network for instance. [ 956.643155] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.645044] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2db09292-d901-42e0-afa0-7144d850a321 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.651486] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47640088-e4c5-487b-b891-8a1eea29a7fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.686241] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb39c90-f671-4127-a40a-cd4ebbdc04d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.698431] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26c5acf-a320-4103-894f-c06648e3bcfb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.714884] env[62000]: DEBUG nova.compute.provider_tree [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.793442] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.793628] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.793741] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 956.890568] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882669, 'name': ReconfigVM_Task, 'duration_secs': 0.167462} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.890874] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 956.891468] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4ff8d8e3-6da4-443a-99bf-8073ec81cb48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.898380] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 956.898380] env[62000]: value = "task-882670" [ 956.898380] env[62000]: _type = "Task" [ 956.898380] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.905953] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882670, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.033283] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.134587] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Received event network-changed-e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 957.134929] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Refreshing instance network info cache due to event network-changed-e041eaad-7b03-44b1-be56-0c3871c66ecc. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 957.135246] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Acquiring lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.135446] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Acquired lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.135618] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Refreshing network info cache for port e041eaad-7b03-44b1-be56-0c3871c66ecc {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 957.147251] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.185923] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.217918] env[62000]: DEBUG nova.scheduler.client.report [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.325919] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 957.409253] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882670, 'name': Rename_Task, 'duration_secs': 0.172334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.409900] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 957.410081] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5e5dcbfc-e6de-4236-abcc-bc554d3c8a0e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.417993] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 957.417993] env[62000]: value = "task-882671" [ 957.417993] env[62000]: _type = "Task" [ 957.417993] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.432171] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882671, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.477140] env[62000]: DEBUG nova.network.neutron [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Updating instance_info_cache with network_info: [{"id": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "address": "fa:16:3e:5b:5c:11", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb11497c-0f", "ovs_interfaceid": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.649503] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.723830] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.724462] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 957.729520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.625s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.730950] env[62000]: INFO nova.compute.claims [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 957.901932] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updated VIF entry in instance network info cache for port e041eaad-7b03-44b1-be56-0c3871c66ecc. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 957.902374] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updating instance_info_cache with network_info: [{"id": "e041eaad-7b03-44b1-be56-0c3871c66ecc", "address": "fa:16:3e:4c:38:15", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": null, "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tape041eaad-7b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.928358] env[62000]: DEBUG oslo_vmware.api [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882671, 'name': PowerOnVM_Task, 'duration_secs': 0.472217} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.928635] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 957.928848] env[62000]: DEBUG nova.compute.manager [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.929641] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa79134c-dfe3-47d6-990d-b2d42095b0b1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.979868] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.980289] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Instance network_info: |[{"id": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "address": "fa:16:3e:5b:5c:11", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb11497c-0f", "ovs_interfaceid": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 957.980671] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.980891] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.981114] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.981302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.981470] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.983431] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:5c:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb224918-e294-4b76-80f9-2fa0031b7dc2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'fb11497c-0f1d-4ff6-baf0-970f2e80bc28', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.991173] env[62000]: DEBUG oslo.service.loopingcall [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.991620] env[62000]: INFO nova.compute.manager [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Terminating instance [ 957.993281] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 957.993841] env[62000]: DEBUG nova.compute.manager [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 957.994044] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 957.994272] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-affea930-66a9-4168-9467-553bcf75ccb4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.009905] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ae0100-40d3-4cec-b22b-9ef3f3eddc32 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.019558] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 958.020812] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-752fe4a6-0294-4c68-a61d-0643e2f011b4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.022257] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.022257] env[62000]: value = "task-882672" [ 958.022257] env[62000]: _type = "Task" [ 958.022257] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.028138] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 958.028138] env[62000]: value = "task-882673" [ 958.028138] env[62000]: _type = "Task" [ 958.028138] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.039473] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.149810] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.235488] env[62000]: DEBUG nova.compute.utils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 958.239487] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 958.239671] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 958.298067] env[62000]: DEBUG nova.policy [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 958.405517] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Releasing lock "refresh_cache-0cbc29c0-b817-4960-9a8d-5fef0b439d58" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.405904] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Received event network-vif-deleted-65be2bd7-e84a-4cc8-b937-6140dc13b2e4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.406156] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Received event network-vif-plugged-fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.406390] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Acquiring lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.406629] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.406808] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.407006] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] No waiting events found dispatching network-vif-plugged-fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 958.407250] env[62000]: WARNING nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Received unexpected event network-vif-plugged-fb11497c-0f1d-4ff6-baf0-970f2e80bc28 for instance with vm_state building and task_state spawning. [ 958.411141] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Received event network-changed-fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 958.411260] env[62000]: DEBUG nova.compute.manager [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Refreshing instance network info cache due to event network-changed-fb11497c-0f1d-4ff6-baf0-970f2e80bc28. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 958.411434] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Acquiring lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.411710] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Acquired lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.411933] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Refreshing network info cache for port fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 958.448176] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.534179] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882672, 'name': CreateVM_Task, 'duration_secs': 0.467257} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.537171] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 958.537860] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.538046] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.538373] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.538942] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f47af824-25d0-49e6-af58-dc37d569a6df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.544078] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882673, 'name': PowerOffVM_Task, 'duration_secs': 0.263152} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.545215] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 958.545400] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 958.545690] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 958.545690] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237ca05-8277-e215-b93f-a80a7110f894" [ 958.545690] env[62000]: _type = "Task" [ 958.545690] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.545876] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-64950dd7-c93b-43bf-90a7-13275104387f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.557010] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237ca05-8277-e215-b93f-a80a7110f894, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.630165] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 958.630633] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 958.630732] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleting the datastore file [datastore2] a0e2344f-32ec-4fb1-ba7c-99e18f422923 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 958.630932] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bba1264a-d7b6-4c0f-bafd-d1a8240fc327 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.639980] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 958.639980] env[62000]: value = "task-882675" [ 958.639980] env[62000]: _type = "Task" [ 958.639980] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.652259] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Successfully created port: 27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 958.660425] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.661211] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882675, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.746423] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 958.971628] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80a95924-e070-429e-8e2b-9b605aec0685 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.982461] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93042588-77dc-436f-bdec-e756aed24d26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.016395] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84155a52-4c99-4fc1-b28d-e8cd9c518281 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.025330] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a253822d-d67e-4864-91ba-aafb76eaba13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.042633] env[62000]: DEBUG nova.compute.provider_tree [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.057983] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237ca05-8277-e215-b93f-a80a7110f894, 'name': SearchDatastore_Task, 'duration_secs': 0.021786} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.059062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.059349] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.059595] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.059746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.059955] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.060483] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1c332a9a-cfc1-4f8c-a825-5f080cee6aa8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.075616] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.075818] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 959.076855] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cfab5ca-512f-4d17-b770-0cad855be2cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.083368] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 959.083368] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52af7382-77f8-8f75-a1aa-bff56c9cb9a7" [ 959.083368] env[62000]: _type = "Task" [ 959.083368] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.094497] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52af7382-77f8-8f75-a1aa-bff56c9cb9a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.162138] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.162460] env[62000]: DEBUG oslo_vmware.api [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882675, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.27858} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.163585] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 959.163843] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 959.164090] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 959.164326] env[62000]: INFO nova.compute.manager [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Took 1.17 seconds to destroy the instance on the hypervisor. [ 959.164593] env[62000]: DEBUG oslo.service.loopingcall [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 959.164793] env[62000]: DEBUG nova.compute.manager [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 959.164904] env[62000]: DEBUG nova.network.neutron [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 959.345076] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Updated VIF entry in instance network info cache for port fb11497c-0f1d-4ff6-baf0-970f2e80bc28. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 959.345582] env[62000]: DEBUG nova.network.neutron [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Updating instance_info_cache with network_info: [{"id": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "address": "fa:16:3e:5b:5c:11", "network": {"id": "28fb2f90-52ec-4639-af79-d57f24ad67d0", "bridge": "br-int", "label": "tempest-ImagesTestJSON-392228401-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d14aec090f2a4abc8078df1fe4428bbb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb224918-e294-4b76-80f9-2fa0031b7dc2", "external-id": "nsx-vlan-transportzone-876", "segmentation_id": 876, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapfb11497c-0f", "ovs_interfaceid": "fb11497c-0f1d-4ff6-baf0-970f2e80bc28", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.366184] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.366184] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.546701] env[62000]: DEBUG nova.scheduler.client.report [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 959.596239] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52af7382-77f8-8f75-a1aa-bff56c9cb9a7, 'name': SearchDatastore_Task, 'duration_secs': 0.017846} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.597341] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-43e3dfbb-1bf3-40ba-9429-daacf784cb70 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.603065] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 959.603065] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a31c5e-315e-bb16-3f98-74b21da1e8c3" [ 959.603065] env[62000]: _type = "Task" [ 959.603065] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.614161] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a31c5e-315e-bb16-3f98-74b21da1e8c3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.651636] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.729863] env[62000]: DEBUG nova.compute.manager [req-057b829e-37bc-4639-b8d3-0345bf9fbefd req-4b443229-941d-4f73-8e50-5bd7edb865a3 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Received event network-vif-deleted-73dc270a-70b4-4343-a230-1102c85ce9e6 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 959.730176] env[62000]: INFO nova.compute.manager [req-057b829e-37bc-4639-b8d3-0345bf9fbefd req-4b443229-941d-4f73-8e50-5bd7edb865a3 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Neutron deleted interface 73dc270a-70b4-4343-a230-1102c85ce9e6; detaching it from the instance and deleting it from the info cache [ 959.730380] env[62000]: DEBUG nova.network.neutron [req-057b829e-37bc-4639-b8d3-0345bf9fbefd req-4b443229-941d-4f73-8e50-5bd7edb865a3 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.757391] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 959.783697] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 959.783978] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 959.784197] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 959.784423] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 959.784602] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 959.784775] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 959.784999] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 959.785209] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 959.785384] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 959.785552] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 959.785729] env[62000]: DEBUG nova.virt.hardware [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 959.786630] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e56f80d7-24ad-49ea-bbf8-ccd8ff260108 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.795053] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa13481-f50a-4a8c-8676-b3ebd979eff8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.848073] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c9b9747-b0b2-461f-b0bc-c58678ddf140 req-6256a099-b316-458b-8ffc-990fd09844eb service nova] Releasing lock "refresh_cache-b6579b75-42dc-473d-8b66-24aa39f9e4ff" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.868274] env[62000]: INFO nova.compute.manager [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Detaching volume 42495829-8539-4f47-9949-eb04eac3dc16 [ 959.905424] env[62000]: INFO nova.virt.block_device [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Attempting to driver detach volume 42495829-8539-4f47-9949-eb04eac3dc16 from mountpoint /dev/sdb [ 959.905657] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 959.905847] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 959.906776] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16d8a8e9-1b10-4e9c-8209-fc2c272a2549 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.931969] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-453ac00c-47b6-48ad-a5ab-9038eff1ba72 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.940804] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bf43220-65e3-4fa9-9df2-710df0c10a8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.963714] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ce5828-a3f4-41fd-8767-dcb890141f86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.980341] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] The volume has not been displaced from its original location: [datastore2] volume-42495829-8539-4f47-9949-eb04eac3dc16/volume-42495829-8539-4f47-9949-eb04eac3dc16.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 959.985720] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfiguring VM instance instance-0000004e to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 959.986045] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d5905452-69b7-47d9-a4fb-b1d197871c2a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.007045] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 960.007045] env[62000]: value = "task-882676" [ 960.007045] env[62000]: _type = "Task" [ 960.007045] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.015503] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882676, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.054008] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.054577] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 960.057227] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 4.520s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.057398] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.057548] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 960.058736] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.753s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.059012] env[62000]: DEBUG nova.objects.instance [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'resources' on Instance uuid 0cbc29c0-b817-4960-9a8d-5fef0b439d58 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.060701] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f1d879b-aa4a-484f-9586-8d1a96d53060 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.069946] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6630bac1-e9f4-4cfc-a8a8-15038f4b760d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.086235] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-908d434b-5d1c-426d-97dd-ebb5fc0aa632 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.094324] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8a80594-8d2b-42f1-b552-3dd0eb67dd2c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.129662] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180140MB free_disk=52GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 960.129892] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.141100] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52a31c5e-315e-bb16-3f98-74b21da1e8c3, 'name': SearchDatastore_Task, 'duration_secs': 0.028767} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.141468] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.141754] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] b6579b75-42dc-473d-8b66-24aa39f9e4ff/b6579b75-42dc-473d-8b66-24aa39f9e4ff.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 960.142061] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-97a56f23-2905-4c68-9f9c-5f52b6d37d58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.153782] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.155221] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 960.155221] env[62000]: value = "task-882677" [ 960.155221] env[62000]: _type = "Task" [ 960.155221] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.163725] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.211043] env[62000]: DEBUG nova.network.neutron [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.235026] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed68f599-7cc2-4163-a48b-c838c7fd1d3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.244700] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237a80b2-a145-4fa5-9c46-df53f0093689 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.280809] env[62000]: DEBUG nova.compute.manager [req-057b829e-37bc-4639-b8d3-0345bf9fbefd req-4b443229-941d-4f73-8e50-5bd7edb865a3 service nova] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Detach interface failed, port_id=73dc270a-70b4-4343-a230-1102c85ce9e6, reason: Instance a0e2344f-32ec-4fb1-ba7c-99e18f422923 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 960.466771] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Successfully updated port: 27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 960.520019] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882676, 'name': ReconfigVM_Task, 'duration_secs': 0.303198} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.520426] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Reconfigured VM instance instance-0000004e to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 960.525316] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-246ae48c-ad30-4de3-b65c-da210e7d673d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.546557] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 960.546557] env[62000]: value = "task-882678" [ 960.546557] env[62000]: _type = "Task" [ 960.546557] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.563668] env[62000]: DEBUG nova.compute.utils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 960.565270] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882678, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.565598] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 960.565768] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 960.569034] env[62000]: DEBUG nova.objects.instance [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'numa_topology' on Instance uuid 0cbc29c0-b817-4960-9a8d-5fef0b439d58 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.616472] env[62000]: DEBUG nova.policy [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e4770851cc3b4ad8aaa6866a18ee1359', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '41e10f7d4f964f2795dc629721802880', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 960.658944] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.669689] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882677, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.713651] env[62000]: INFO nova.compute.manager [-] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Took 1.55 seconds to deallocate network for instance. [ 960.973956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.973956] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.973956] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 961.059331] env[62000]: DEBUG oslo_vmware.api [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882678, 'name': ReconfigVM_Task, 'duration_secs': 0.185008} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.059331] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201615', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'name': 'volume-42495829-8539-4f47-9949-eb04eac3dc16', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'd2f87d82-af5e-4b17-959e-40865a8852b3', 'attached_at': '', 'detached_at': '', 'volume_id': '42495829-8539-4f47-9949-eb04eac3dc16', 'serial': '42495829-8539-4f47-9949-eb04eac3dc16'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 961.074020] env[62000]: DEBUG nova.objects.base [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Object Instance<0cbc29c0-b817-4960-9a8d-5fef0b439d58> lazy-loaded attributes: resources,numa_topology {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 961.074020] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 961.141797] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Successfully created port: a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 961.157412] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.167788] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.554594} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.170598] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] b6579b75-42dc-473d-8b66-24aa39f9e4ff/b6579b75-42dc-473d-8b66-24aa39f9e4ff.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 961.170734] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 961.171066] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ec9777c-33b2-45a5-a492-8ddc2f08eadc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.178712] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 961.178712] env[62000]: value = "task-882679" [ 961.178712] env[62000]: _type = "Task" [ 961.178712] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.189698] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882679, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.222921] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.300073] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3e0d9d0-c409-40e8-9d8f-a2655f74c023 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.308381] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2771ef-c013-4dad-8f65-f44d8a6d529c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.342734] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36dc0eb2-be3f-403f-95fb-b6fd565bcc85 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.351455] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d4b6957-f468-4511-836b-512f16c9f492 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.366626] env[62000]: DEBUG nova.compute.provider_tree [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.503092] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 961.608266] env[62000]: DEBUG nova.objects.instance [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'flavor' on Instance uuid d2f87d82-af5e-4b17-959e-40865a8852b3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.636250] env[62000]: DEBUG nova.network.neutron [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Updating instance_info_cache with network_info: [{"id": "27537c6a-a039-4b84-877b-e91b648742e0", "address": "fa:16:3e:d0:34:69", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27537c6a-a0", "ovs_interfaceid": "27537c6a-a039-4b84-877b-e91b648742e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 961.657497] env[62000]: DEBUG oslo_vmware.api [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882667, 'name': ReconfigVM_Task, 'duration_secs': 5.757099} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.657768] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 961.657990] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Reconfigured VM to detach interface {{(pid=62000) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1984}} [ 961.689124] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882679, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.264648} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.691587] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 961.692499] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c5c9a57-b311-43b8-b7b2-faa4188ea162 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.716875] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] b6579b75-42dc-473d-8b66-24aa39f9e4ff/b6579b75-42dc-473d-8b66-24aa39f9e4ff.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 961.717230] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-85f99882-cde7-4ded-bbf9-331390a19a6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.741269] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 961.741269] env[62000]: value = "task-882680" [ 961.741269] env[62000]: _type = "Task" [ 961.741269] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.750385] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882680, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.759104] env[62000]: DEBUG nova.compute.manager [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Received event network-vif-plugged-27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.759354] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Acquiring lock "23366d62-294b-47c8-adc6-80e854f452ef-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.759583] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Lock "23366d62-294b-47c8-adc6-80e854f452ef-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.759716] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Lock "23366d62-294b-47c8-adc6-80e854f452ef-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.759971] env[62000]: DEBUG nova.compute.manager [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] No waiting events found dispatching network-vif-plugged-27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 961.760304] env[62000]: WARNING nova.compute.manager [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Received unexpected event network-vif-plugged-27537c6a-a039-4b84-877b-e91b648742e0 for instance with vm_state building and task_state spawning. [ 961.760502] env[62000]: DEBUG nova.compute.manager [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Received event network-changed-27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 961.760673] env[62000]: DEBUG nova.compute.manager [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Refreshing instance network info cache due to event network-changed-27537c6a-a039-4b84-877b-e91b648742e0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 961.760854] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Acquiring lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.870417] env[62000]: DEBUG nova.scheduler.client.report [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.086042] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 962.111047] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 962.111342] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 962.111506] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 962.111820] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 962.111910] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 962.111985] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 962.112277] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 962.112467] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 962.112645] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 962.112833] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 962.112986] env[62000]: DEBUG nova.virt.hardware [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 962.115710] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf44f430-1c05-4820-9aec-da1ba1a81ca5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.126077] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0d5443c-333f-4c1b-a3e3-bcb90ab784fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.144012] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 962.144371] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Instance network_info: |[{"id": "27537c6a-a039-4b84-877b-e91b648742e0", "address": "fa:16:3e:d0:34:69", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27537c6a-a0", "ovs_interfaceid": "27537c6a-a039-4b84-877b-e91b648742e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 962.144976] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Acquired lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.145258] env[62000]: DEBUG nova.network.neutron [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Refreshing network info cache for port 27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 962.146692] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:34:69', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '27537c6a-a039-4b84-877b-e91b648742e0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 962.154195] env[62000]: DEBUG oslo.service.loopingcall [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.155143] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 962.155442] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-53ebea05-f517-4008-84cd-307ea1082e19 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.178318] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 962.178318] env[62000]: value = "task-882681" [ 962.178318] env[62000]: _type = "Task" [ 962.178318] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.186922] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882681, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.252512] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882680, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.376953] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.319s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.379560] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.346s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.379878] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.382257] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 3.934s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.382466] env[62000]: DEBUG nova.objects.instance [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62000) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 962.403788] env[62000]: INFO nova.scheduler.client.report [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocations for instance dfa26fab-40ff-4409-bb73-79aa555ab225 [ 962.621474] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52229235-e178-4885-b946-432d9b2c61f4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.256s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.688654] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882681, 'name': CreateVM_Task, 'duration_secs': 0.306976} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.688832] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 962.689531] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.689697] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.690054] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 962.690313] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d86e7123-1034-4c85-ab5d-4309123badb4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.694977] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 962.694977] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b7447e-1106-7d2f-d13d-12c64d484f1c" [ 962.694977] env[62000]: _type = "Task" [ 962.694977] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.705240] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b7447e-1106-7d2f-d13d-12c64d484f1c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.751608] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882680, 'name': ReconfigVM_Task, 'duration_secs': 0.549707} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.751924] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Reconfigured VM instance instance-0000005a to attach disk [datastore2] b6579b75-42dc-473d-8b66-24aa39f9e4ff/b6579b75-42dc-473d-8b66-24aa39f9e4ff.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 962.752610] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9856ca7c-d526-4764-a6b1-ec9cb2790f5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.761230] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 962.761230] env[62000]: value = "task-882682" [ 962.761230] env[62000]: _type = "Task" [ 962.761230] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.769781] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882682, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.785401] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Successfully updated port: a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 962.893524] env[62000]: DEBUG oslo_concurrency.lockutils [None req-598f9941-654b-49bf-819a-495b4d4183a6 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 29.470s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.894492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.709s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.894745] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.894987] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.895181] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.896890] env[62000]: INFO nova.compute.manager [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Terminating instance [ 962.899148] env[62000]: DEBUG nova.compute.manager [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 962.899278] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 962.899532] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bb7364f6-7835-4e7c-87e4-966c8843ca79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.902213] env[62000]: DEBUG nova.network.neutron [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Updated VIF entry in instance network info cache for port 27537c6a-a039-4b84-877b-e91b648742e0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 962.902539] env[62000]: DEBUG nova.network.neutron [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Updating instance_info_cache with network_info: [{"id": "27537c6a-a039-4b84-877b-e91b648742e0", "address": "fa:16:3e:d0:34:69", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap27537c6a-a0", "ovs_interfaceid": "27537c6a-a039-4b84-877b-e91b648742e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.913633] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2cb19267-cce2-4ca4-8a14-39b80b22ad90 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "dfa26fab-40ff-4409-bb73-79aa555ab225" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.332s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.919347] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2bc9f7-45c1-4336-b60c-711e78647968 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.954612] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0cbc29c0-b817-4960-9a8d-5fef0b439d58 could not be found. [ 962.954824] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 962.955016] env[62000]: INFO nova.compute.manager [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Took 0.06 seconds to destroy the instance on the hypervisor. [ 962.955267] env[62000]: DEBUG oslo.service.loopingcall [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 962.955517] env[62000]: DEBUG nova.compute.manager [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 962.955611] env[62000]: DEBUG nova.network.neutron [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 963.002355] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.002535] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquired lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.002713] env[62000]: DEBUG nova.network.neutron [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.210866] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b7447e-1106-7d2f-d13d-12c64d484f1c, 'name': SearchDatastore_Task, 'duration_secs': 0.012155} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.211932] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.213618] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 963.214056] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.214056] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.214233] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 963.214644] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e48ce0d5-7883-48ea-83d8-5c815254fbb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.224766] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 963.224964] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 963.225712] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-37fdc7c7-b306-4d8a-bcb3-2811e8e99ab6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.235066] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 963.235066] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5283501b-d829-aa8d-68d4-8aab9ef88812" [ 963.235066] env[62000]: _type = "Task" [ 963.235066] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.243441] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5283501b-d829-aa8d-68d4-8aab9ef88812, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.271350] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882682, 'name': Rename_Task, 'duration_secs': 0.14914} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.271621] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 963.271866] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c0b182d-7403-4ea2-b278-28c0f3a75a8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.280710] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 963.280710] env[62000]: value = "task-882683" [ 963.280710] env[62000]: _type = "Task" [ 963.280710] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.292424] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.292424] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.292424] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 963.292424] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882683, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.395119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-43e299e9-8219-4f50-b558-16e57c96e8d4 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.013s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.396575] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 3.267s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.404873] env[62000]: DEBUG oslo_concurrency.lockutils [req-47cf3800-874e-4c0f-936c-a70b575b6bd1 req-308cf688-d3a1-48c7-9f14-ade0a3fd04ac service nova] Releasing lock "refresh_cache-23366d62-294b-47c8-adc6-80e854f452ef" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.505529] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.505926] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.506242] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.506512] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.506756] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.511702] env[62000]: INFO nova.compute.manager [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Terminating instance [ 963.514032] env[62000]: DEBUG nova.compute.manager [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.514298] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 963.515767] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebcdd4e2-7685-4586-ac21-b2c35a796408 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.525567] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 963.525818] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b5e84e6a-2c7e-453b-951d-cda0ed4bd319 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.532847] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 963.532847] env[62000]: value = "task-882684" [ 963.532847] env[62000]: _type = "Task" [ 963.532847] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.541096] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882684, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.697452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.697452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.697452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.697452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.697452] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.698346] env[62000]: INFO nova.compute.manager [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Terminating instance [ 963.700666] env[62000]: DEBUG nova.compute.manager [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.701019] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 963.702119] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f217781d-055b-4c49-b6de-bd631a2598dd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.711193] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 963.711581] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-400dfd4f-1525-4f9e-9226-712068cdd52f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.720048] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 963.720048] env[62000]: value = "task-882685" [ 963.720048] env[62000]: _type = "Task" [ 963.720048] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.727340] env[62000]: DEBUG nova.network.neutron [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.728841] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.743948] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5283501b-d829-aa8d-68d4-8aab9ef88812, 'name': SearchDatastore_Task, 'duration_secs': 0.010478} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.745599] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bc4566ef-f73d-45d9-93d0-655d154220e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.754016] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 963.754016] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cd6eff-0172-d604-757a-00d1bc538365" [ 963.754016] env[62000]: _type = "Task" [ 963.754016] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.766068] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cd6eff-0172-d604-757a-00d1bc538365, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.773516] env[62000]: INFO nova.network.neutron [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Port 68fd0d34-536e-4504-a837-62eca9533e5f from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 963.773941] env[62000]: DEBUG nova.network.neutron [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [{"id": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "address": "fa:16:3e:9a:5f:74", "network": {"id": "e21c3795-1c6b-42ef-af81-e113912fa80d", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1749839473-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "747546b09fd04a41b9c2df860a699186", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a15de394-0367-4921-a5c1-6ac8615e3283", "external-id": "nsx-vlan-transportzone-13", "segmentation_id": 13, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1e2f48d0-f8", "ovs_interfaceid": "1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.792446] env[62000]: DEBUG oslo_vmware.api [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882683, 'name': PowerOnVM_Task, 'duration_secs': 0.486958} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.795335] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 963.795632] env[62000]: INFO nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Took 8.15 seconds to spawn the instance on the hypervisor. [ 963.795898] env[62000]: DEBUG nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 963.797132] env[62000]: DEBUG nova.compute.manager [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Received event network-vif-plugged-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.797403] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Acquiring lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.797650] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.797886] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.798145] env[62000]: DEBUG nova.compute.manager [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] No waiting events found dispatching network-vif-plugged-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 963.798487] env[62000]: WARNING nova.compute.manager [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Received unexpected event network-vif-plugged-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe for instance with vm_state building and task_state spawning. [ 963.798723] env[62000]: DEBUG nova.compute.manager [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Received event network-changed-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 963.798942] env[62000]: DEBUG nova.compute.manager [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Refreshing instance network info cache due to event network-changed-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 963.799189] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Acquiring lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 963.802568] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d94f5a-d14c-4a0d-b2df-51055bdbb49a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.835537] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 963.973411] env[62000]: DEBUG nova.network.neutron [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Updating instance_info_cache with network_info: [{"id": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "address": "fa:16:3e:fa:a8:32", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7dcf91b-15", "ovs_interfaceid": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.042631] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882684, 'name': PowerOffVM_Task, 'duration_secs': 0.164569} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.042894] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 964.043131] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 964.043389] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dcde5184-901b-4cb8-9e9e-f2599fd38a97 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.114885] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 964.114885] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 964.114885] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleting the datastore file [datastore2] 94605ab8-8167-430e-b1cd-c8f51e50d8b9 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.115349] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f27bdbf8-0e33-405e-b27d-55c64f1d02e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.123526] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 964.123526] env[62000]: value = "task-882687" [ 964.123526] env[62000]: _type = "Task" [ 964.123526] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.133706] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882687, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.228261] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882685, 'name': PowerOffVM_Task, 'duration_secs': 0.181246} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.228589] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 964.228757] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 964.229382] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-62006bc8-76a0-40e8-b36b-0cc54d682d8b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.230755] env[62000]: INFO nova.compute.manager [-] [instance: 0cbc29c0-b817-4960-9a8d-5fef0b439d58] Took 1.28 seconds to deallocate network for instance. [ 964.264532] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cd6eff-0172-d604-757a-00d1bc538365, 'name': SearchDatastore_Task, 'duration_secs': 0.008963} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.265513] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.265792] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 23366d62-294b-47c8-adc6-80e854f452ef/23366d62-294b-47c8-adc6-80e854f452ef.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 964.266072] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d492f05-bfb9-496d-8829-d753d50a919a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.272777] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 964.272777] env[62000]: value = "task-882689" [ 964.272777] env[62000]: _type = "Task" [ 964.272777] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.276455] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Releasing lock "refresh_cache-94605ab8-8167-430e-b1cd-c8f51e50d8b9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.283713] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882689, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.306499] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 964.306755] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 964.306943] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] d2f87d82-af5e-4b17-959e-40865a8852b3 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.307226] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ebae8946-0244-41a6-b61e-61da0f9869e0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.313843] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 964.313843] env[62000]: value = "task-882690" [ 964.313843] env[62000]: _type = "Task" [ 964.313843] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.326614] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882690, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.327097] env[62000]: INFO nova.compute.manager [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Took 13.05 seconds to build instance. [ 964.432221] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432344] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 66edb906-f936-4f57-833d-224f36af109e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432474] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 3780b22d-c360-4433-9f6b-9d5d1f14b525 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432629] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 44a54bd7-a35c-49ed-85ed-346830cee6ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432725] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 72a5bf70-dc6e-4887-abb8-8fbad64bb065 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432840] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance d2f87d82-af5e-4b17-959e-40865a8852b3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.432966] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 94605ab8-8167-430e-b1cd-c8f51e50d8b9 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433188] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance a0e2344f-32ec-4fb1-ba7c-99e18f422923 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 964.433337] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 8a11689f-fc00-43f8-9215-8d81daa84400 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433449] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 3878579f-6435-4fe3-9f8c-8461d8ac57ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433561] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance ea58da8c-e4ac-4863-942d-6294cf04e6f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433693] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance b6579b75-42dc-473d-8b66-24aa39f9e4ff actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433807] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 23366d62-294b-47c8-adc6-80e854f452ef actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.433894] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 0e66a948-61d5-4991-99bd-374db7eaaf2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 964.434128] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 964.434281] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3008MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 964.475640] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.476155] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Instance network_info: |[{"id": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "address": "fa:16:3e:fa:a8:32", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7dcf91b-15", "ovs_interfaceid": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 964.476314] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Acquired lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 964.476501] env[62000]: DEBUG nova.network.neutron [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Refreshing network info cache for port a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 964.477847] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:a8:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e365f3b9-706b-4fa2-8f95-ae51b35ab011', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a7dcf91b-1563-4262-86e8-ee28ac8d0bbe', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 964.485540] env[62000]: DEBUG oslo.service.loopingcall [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.491116] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 964.492173] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0751bad1-599e-4a61-afd8-dfe24ac0b2a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.516380] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 964.516380] env[62000]: value = "task-882691" [ 964.516380] env[62000]: _type = "Task" [ 964.516380] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.527655] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882691, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.645465] env[62000]: DEBUG oslo_vmware.api [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882687, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.164545} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.646006] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.646546] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 964.646786] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 964.646981] env[62000]: INFO nova.compute.manager [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Took 1.13 seconds to destroy the instance on the hypervisor. [ 964.647251] env[62000]: DEBUG oslo.service.loopingcall [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.647502] env[62000]: DEBUG nova.compute.manager [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 964.647600] env[62000]: DEBUG nova.network.neutron [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 964.693046] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a34fa890-6bba-48a3-b7b3-159dac7bae9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.704581] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83778a21-4bbd-48f4-863d-3f15e5710991 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.749962] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65679c98-3091-4221-b7ce-dd3a09d82483 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.760184] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b034e98c-25d1-417c-a097-aea9e46027a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.779579] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 964.783145] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb4f11fb-21f7-4b71-8d86-1b5c5814f03c tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "interface-94605ab8-8167-430e-b1cd-c8f51e50d8b9-68fd0d34-536e-4504-a837-62eca9533e5f" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.720s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.793735] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882689, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51452} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.796541] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 23366d62-294b-47c8-adc6-80e854f452ef/23366d62-294b-47c8-adc6-80e854f452ef.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 964.796765] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 964.797284] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-48a050b6-25f4-4df0-a2e9-c3515e38c6cc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.804284] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 964.804284] env[62000]: value = "task-882692" [ 964.804284] env[62000]: _type = "Task" [ 964.804284] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.812746] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882692, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.824200] env[62000]: DEBUG oslo_vmware.api [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882690, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.158208} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.824946] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.825176] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 964.825521] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 964.825647] env[62000]: INFO nova.compute.manager [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Took 1.12 seconds to destroy the instance on the hypervisor. [ 964.825815] env[62000]: DEBUG oslo.service.loopingcall [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.827557] env[62000]: DEBUG nova.compute.manager [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 964.827557] env[62000]: DEBUG nova.network.neutron [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 964.828754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-eba87c0a-ba4f-41cf-ac82-1b7596988215 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.559s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.023851] env[62000]: DEBUG nova.network.neutron [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Updated VIF entry in instance network info cache for port a7dcf91b-1563-4262-86e8-ee28ac8d0bbe. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 965.024252] env[62000]: DEBUG nova.network.neutron [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Updating instance_info_cache with network_info: [{"id": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "address": "fa:16:3e:fa:a8:32", "network": {"id": "7c58b319-49fc-47c3-a1a6-40df03c4fb02", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1252932514-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "41e10f7d4f964f2795dc629721802880", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e365f3b9-706b-4fa2-8f95-ae51b35ab011", "external-id": "nsx-vlan-transportzone-154", "segmentation_id": 154, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa7dcf91b-15", "ovs_interfaceid": "a7dcf91b-1563-4262-86e8-ee28ac8d0bbe", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.028830] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882691, 'name': CreateVM_Task, 'duration_secs': 0.433855} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.029212] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 965.029859] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.030029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.030351] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 965.030606] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22a0eb52-a861-4e65-80b9-fc4704390275 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.035093] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 965.035093] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527253a6-ad33-1196-d417-1ab1dfdecc57" [ 965.035093] env[62000]: _type = "Task" [ 965.035093] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.044013] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527253a6-ad33-1196-d417-1ab1dfdecc57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.283269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bf96f73c-5d41-49a7-b718-b55e06bb3f1a tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "0cbc29c0-b817-4960-9a8d-5fef0b439d58" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.389s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.285296] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.317380] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882692, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08193} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.317643] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 965.318445] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d171fe5-39d5-410b-899a-104af25f8cd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.340281] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 23366d62-294b-47c8-adc6-80e854f452ef/23366d62-294b-47c8-adc6-80e854f452ef.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 965.340560] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bff88939-e868-452e-9939-0b1ea50b0b87 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.360895] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 965.360895] env[62000]: value = "task-882693" [ 965.360895] env[62000]: _type = "Task" [ 965.360895] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.369850] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882693, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.529942] env[62000]: DEBUG oslo_concurrency.lockutils [req-aacf0068-c014-4a44-8e82-dbf36a0cdcc8 req-ab30caf2-b4fd-4dea-b071-13c01f7bd803 service nova] Releasing lock "refresh_cache-0e66a948-61d5-4991-99bd-374db7eaaf2a" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.545896] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527253a6-ad33-1196-d417-1ab1dfdecc57, 'name': SearchDatastore_Task, 'duration_secs': 0.008905} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.546206] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 965.546441] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 965.546684] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.546832] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.547029] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.547298] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4248b810-6d3d-4759-94f6-b2a357bd7d11 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.556204] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.556364] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 965.557102] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e08b50d-cf50-4add-b7aa-ac30c5ee387f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.562488] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 965.562488] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dd0d11-7f29-7cd4-b423-4db19a9064e7" [ 965.562488] env[62000]: _type = "Task" [ 965.562488] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.572512] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dd0d11-7f29-7cd4-b423-4db19a9064e7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.621017] env[62000]: DEBUG nova.compute.manager [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 965.621017] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e6c0bf5-bd10-4e1c-abd7-654ad3996c4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.792945] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 965.793143] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.397s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.794837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.571s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.794837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.831505] env[62000]: INFO nova.scheduler.client.report [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted allocations for instance a0e2344f-32ec-4fb1-ba7c-99e18f422923 [ 965.836320] env[62000]: DEBUG nova.compute.manager [req-2344fb67-7548-4400-8ed6-ebf6e9da1a76 req-5508ca36-af20-4ba3-8c38-8e45a152c424 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Received event network-vif-deleted-919b3f5d-af5d-45e1-896f-ac0ea1309017 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 965.836320] env[62000]: INFO nova.compute.manager [req-2344fb67-7548-4400-8ed6-ebf6e9da1a76 req-5508ca36-af20-4ba3-8c38-8e45a152c424 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Neutron deleted interface 919b3f5d-af5d-45e1-896f-ac0ea1309017; detaching it from the instance and deleting it from the info cache [ 965.836320] env[62000]: DEBUG nova.network.neutron [req-2344fb67-7548-4400-8ed6-ebf6e9da1a76 req-5508ca36-af20-4ba3-8c38-8e45a152c424 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.876549] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882693, 'name': ReconfigVM_Task, 'duration_secs': 0.274219} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.876836] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 23366d62-294b-47c8-adc6-80e854f452ef/23366d62-294b-47c8-adc6-80e854f452ef.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 965.877639] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6042800-6b17-48d3-9632-96ebdafa32c6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.885361] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 965.885361] env[62000]: value = "task-882694" [ 965.885361] env[62000]: _type = "Task" [ 965.885361] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.897407] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882694, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.078990] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dd0d11-7f29-7cd4-b423-4db19a9064e7, 'name': SearchDatastore_Task, 'duration_secs': 0.009127} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.080045] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9731a5c-c02e-4484-98c7-c930b666cd3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.087781] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 966.087781] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5286e15a-cce2-4e2e-51fd-8b173d2ab0d6" [ 966.087781] env[62000]: _type = "Task" [ 966.087781] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.096353] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5286e15a-cce2-4e2e-51fd-8b173d2ab0d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.130838] env[62000]: INFO nova.compute.manager [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] instance snapshotting [ 966.136191] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40925c7d-b5ae-4247-9347-202d7fc62c13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.159816] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12db43c4-79b7-4a5e-9685-969350726540 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.174777] env[62000]: DEBUG nova.network.neutron [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.346914] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0adafd12-85ae-4ca4-9d06-0c7c7674fc3a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "a0e2344f-32ec-4fb1-ba7c-99e18f422923" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.366s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.347933] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-761bdd66-9dbc-4186-a03a-98494fc7a648 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.361419] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64610890-f904-494a-8f9c-17d35773ed64 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.371852] env[62000]: DEBUG nova.network.neutron [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.395664] env[62000]: DEBUG nova.compute.manager [req-2344fb67-7548-4400-8ed6-ebf6e9da1a76 req-5508ca36-af20-4ba3-8c38-8e45a152c424 service nova] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Detach interface failed, port_id=919b3f5d-af5d-45e1-896f-ac0ea1309017, reason: Instance d2f87d82-af5e-4b17-959e-40865a8852b3 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 966.405863] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882694, 'name': Rename_Task, 'duration_secs': 0.128898} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.406188] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 966.406447] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e8bf75bd-94b4-4b2a-8044-26f3d0e1e31b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.413049] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 966.413049] env[62000]: value = "task-882695" [ 966.413049] env[62000]: _type = "Task" [ 966.413049] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.421595] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882695, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.599011] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5286e15a-cce2-4e2e-51fd-8b173d2ab0d6, 'name': SearchDatastore_Task, 'duration_secs': 0.010367} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.599439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.599787] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 0e66a948-61d5-4991-99bd-374db7eaaf2a/0e66a948-61d5-4991-99bd-374db7eaaf2a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 966.600077] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0ada57f9-e539-48f3-be05-3d620837ad4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.606829] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 966.606829] env[62000]: value = "task-882696" [ 966.606829] env[62000]: _type = "Task" [ 966.606829] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.614386] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882696, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.670799] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 966.671158] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-887e54f2-7944-4541-bdc4-406e9238c841 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.677356] env[62000]: INFO nova.compute.manager [-] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Took 1.85 seconds to deallocate network for instance. [ 966.680628] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 966.680628] env[62000]: value = "task-882697" [ 966.680628] env[62000]: _type = "Task" [ 966.680628] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.692733] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882697, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.737913] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.738951] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.873323] env[62000]: INFO nova.compute.manager [-] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Took 2.23 seconds to deallocate network for instance. [ 966.923096] env[62000]: DEBUG oslo_vmware.api [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882695, 'name': PowerOnVM_Task, 'duration_secs': 0.450481} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.923385] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 966.923590] env[62000]: INFO nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Took 7.17 seconds to spawn the instance on the hypervisor. [ 966.923781] env[62000]: DEBUG nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 966.924593] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a998ee9-9fb7-4688-824b-805068d2fc88 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.117516] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882696, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.131385] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.131664] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.191785] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.192096] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.192396] env[62000]: DEBUG nova.objects.instance [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'resources' on Instance uuid d2f87d82-af5e-4b17-959e-40865a8852b3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.193581] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882697, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.241349] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 967.367556] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.367857] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.380342] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.442999] env[62000]: INFO nova.compute.manager [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Took 13.64 seconds to build instance. [ 967.618845] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882696, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.638178} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.619138] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 0e66a948-61d5-4991-99bd-374db7eaaf2a/0e66a948-61d5-4991-99bd-374db7eaaf2a.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 967.619381] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 967.619639] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9323386e-ce11-43a1-ac46-5fa0536e8484 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.625732] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 967.625732] env[62000]: value = "task-882698" [ 967.625732] env[62000]: _type = "Task" [ 967.625732] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.633169] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.639539] env[62000]: DEBUG nova.compute.utils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 967.692174] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882697, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.762772] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 967.869862] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b40b168-c4a8-4df9-b074-51918fbc843c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.873904] env[62000]: INFO nova.compute.manager [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Detaching volume c6426f66-679d-4eb9-9518-08c638760409 [ 967.883629] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94011129-47e8-4590-9c7c-262b3033e270 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.915121] env[62000]: DEBUG nova.compute.manager [req-19df5906-e7b6-450e-b671-7581dde2eade req-e98b760f-0301-4216-895f-78735478f9cd service nova] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Received event network-vif-deleted-1e2f48d0-f8b8-4f72-b5f9-805bbb5bb583 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 967.915978] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fdacf71-5824-42fc-b81f-8be4236433ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.919331] env[62000]: INFO nova.virt.block_device [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Attempting to driver detach volume c6426f66-679d-4eb9-9518-08c638760409 from mountpoint /dev/sdb [ 967.919549] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 967.919739] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201596', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'name': 'volume-c6426f66-679d-4eb9-9518-08c638760409', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3780b22d-c360-4433-9f6b-9d5d1f14b525', 'attached_at': '', 'detached_at': '', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'serial': 'c6426f66-679d-4eb9-9518-08c638760409'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 967.920707] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f41f933-5d84-4429-90fe-7da95ce9e0f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.944162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0ccbe59-f100-4f73-8624-e9f490aa0ca3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.948517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c02d143-bc08-41b7-9de0-ae264fd9643d tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.158s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.949193] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecc82d01-9922-44f4-a395-5564e7345c7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.961755] env[62000]: DEBUG nova.compute.provider_tree [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 967.964908] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0491be69-04c1-4c64-8b29-55a6c5167fcf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.984911] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f824fb-2012-4e3e-8fc0-385a5fdaca49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.999443] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] The volume has not been displaced from its original location: [datastore1] volume-c6426f66-679d-4eb9-9518-08c638760409/volume-c6426f66-679d-4eb9-9518-08c638760409.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 968.004664] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 968.004940] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-45da67fe-ec27-4934-a7c6-4a250c3e3962 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.023508] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 968.023508] env[62000]: value = "task-882699" [ 968.023508] env[62000]: _type = "Task" [ 968.023508] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.031061] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882699, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.135674] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069465} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.136024] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 968.137076] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d5a482-5c6a-4775-ac2f-42c4b8734349 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.150129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.018s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.159046] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 0e66a948-61d5-4991-99bd-374db7eaaf2a/0e66a948-61d5-4991-99bd-374db7eaaf2a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.159370] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c162b42e-ce64-4d6e-9158-6ead06d05708 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.179705] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 968.179705] env[62000]: value = "task-882700" [ 968.179705] env[62000]: _type = "Task" [ 968.179705] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.188692] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882700, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.193379] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882697, 'name': CreateSnapshot_Task, 'duration_secs': 1.040075} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.193624] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 968.194323] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbe4edd-9d2d-4f3e-b689-082030c8ac42 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.373627] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "23366d62-294b-47c8-adc6-80e854f452ef" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.374009] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.374282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "23366d62-294b-47c8-adc6-80e854f452ef-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.374514] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.374741] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.377026] env[62000]: INFO nova.compute.manager [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Terminating instance [ 968.379202] env[62000]: DEBUG nova.compute.manager [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 968.379404] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 968.380248] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b37e607-82b4-467b-9f31-7c88957a1cda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.387822] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 968.388069] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2da5bdb-247b-484b-bb78-52c8ee07a174 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.394164] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 968.394164] env[62000]: value = "task-882701" [ 968.394164] env[62000]: _type = "Task" [ 968.394164] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.401259] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882701, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.465769] env[62000]: DEBUG nova.scheduler.client.report [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 968.535069] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882699, 'name': ReconfigVM_Task, 'duration_secs': 0.373142} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.535069] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 968.538463] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4ccf6f5-c215-43fe-b1b9-64b8faa568aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.554015] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 968.554015] env[62000]: value = "task-882702" [ 968.554015] env[62000]: _type = "Task" [ 968.554015] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.561691] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882702, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.690624] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882700, 'name': ReconfigVM_Task, 'duration_secs': 0.323167} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.692129] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 0e66a948-61d5-4991-99bd-374db7eaaf2a/0e66a948-61d5-4991-99bd-374db7eaaf2a.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.694798] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b7e30027-64e6-4e55-96f3-260440d71fe2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.702115] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 968.702115] env[62000]: value = "task-882703" [ 968.702115] env[62000]: _type = "Task" [ 968.702115] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.712179] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 968.712969] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-72306e06-54da-427e-a07b-abf8fc4e7708 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.719248] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882703, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.724228] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 968.724228] env[62000]: value = "task-882704" [ 968.724228] env[62000]: _type = "Task" [ 968.724228] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.732127] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882704, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.904702] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882701, 'name': PowerOffVM_Task, 'duration_secs': 0.150297} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.904949] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 968.905109] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 968.905370] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2e9074e-a3bb-4aa1-a0e5-5f70a7052cbe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.969261] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 968.969506] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 968.969683] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore2] 23366d62-294b-47c8-adc6-80e854f452ef {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.969974] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f7781b04-86a6-4e10-a5b3-f7ec018bd3e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.972431] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.975045] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.594s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.975045] env[62000]: DEBUG nova.objects.instance [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'resources' on Instance uuid 94605ab8-8167-430e-b1cd-c8f51e50d8b9 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.980819] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 968.980819] env[62000]: value = "task-882706" [ 968.980819] env[62000]: _type = "Task" [ 968.980819] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.990015] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882706, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.992454] env[62000]: INFO nova.scheduler.client.report [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted allocations for instance d2f87d82-af5e-4b17-959e-40865a8852b3 [ 969.065528] env[62000]: DEBUG oslo_vmware.api [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882702, 'name': ReconfigVM_Task, 'duration_secs': 0.131434} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.065844] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201596', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'name': 'volume-c6426f66-679d-4eb9-9518-08c638760409', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3780b22d-c360-4433-9f6b-9d5d1f14b525', 'attached_at': '', 'detached_at': '', 'volume_id': 'c6426f66-679d-4eb9-9518-08c638760409', 'serial': 'c6426f66-679d-4eb9-9518-08c638760409'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 969.197968] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.198322] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.198605] env[62000]: INFO nova.compute.manager [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Attaching volume f21e6dba-4892-4d3a-9758-294a3d9d63f7 to /dev/sdb [ 969.214228] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882703, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.233789] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882704, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.235964] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b3ed9e-d510-4b70-af5f-1f3ec7b8bb7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.241951] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121465f5-2116-4ca2-b552-290ac152e097 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.256267] env[62000]: DEBUG nova.virt.block_device [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating existing volume attachment record: ca5bee9d-8186-4e8b-a055-a14dbf32f1cd {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 969.489628] env[62000]: DEBUG oslo_vmware.api [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882706, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185817} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.489961] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.490216] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 969.490303] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 969.490453] env[62000]: INFO nova.compute.manager [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Took 1.11 seconds to destroy the instance on the hypervisor. [ 969.490697] env[62000]: DEBUG oslo.service.loopingcall [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 969.490893] env[62000]: DEBUG nova.compute.manager [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 969.490989] env[62000]: DEBUG nova.network.neutron [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 969.503877] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8cb594a9-705c-47c3-ace4-3de271ce159f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "d2f87d82-af5e-4b17-959e-40865a8852b3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.808s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.611421] env[62000]: DEBUG nova.objects.instance [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 3780b22d-c360-4433-9f6b-9d5d1f14b525 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.663305] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4de7048a-cb59-4f49-8400-af2d340d9a1d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.672368] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65d6e18c-0ec0-45f4-974d-fca1052333bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.710644] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9953cfdb-5bcb-4a60-93f8-db9b2554816e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.721401] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882703, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.726458] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8806308a-8a6f-4731-9c4b-fda542f3f3d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.739804] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882704, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.748146] env[62000]: DEBUG nova.compute.provider_tree [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 969.918551] env[62000]: DEBUG nova.compute.manager [req-2dd10f33-cc99-452e-a504-4f144ce0c1d0 req-9db6b71f-0d90-4b0f-9100-3db3df70b2a0 service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Received event network-vif-deleted-27537c6a-a039-4b84-877b-e91b648742e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 969.918732] env[62000]: INFO nova.compute.manager [req-2dd10f33-cc99-452e-a504-4f144ce0c1d0 req-9db6b71f-0d90-4b0f-9100-3db3df70b2a0 service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Neutron deleted interface 27537c6a-a039-4b84-877b-e91b648742e0; detaching it from the instance and deleting it from the info cache [ 969.918917] env[62000]: DEBUG nova.network.neutron [req-2dd10f33-cc99-452e-a504-4f144ce0c1d0 req-9db6b71f-0d90-4b0f-9100-3db3df70b2a0 service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.216629] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882703, 'name': Rename_Task, 'duration_secs': 1.153359} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.217275] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 970.217275] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-944ed75e-3b70-48ad-835d-b069fa4269f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.224829] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 970.224829] env[62000]: value = "task-882710" [ 970.224829] env[62000]: _type = "Task" [ 970.224829] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.227881] env[62000]: DEBUG nova.network.neutron [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.234427] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882710, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.243063] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882704, 'name': CloneVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.252505] env[62000]: DEBUG nova.scheduler.client.report [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 970.422291] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c4821f1-3622-45ce-8824-9d9d03e7b18a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.432278] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d5f97d-22eb-40f6-b7e8-91d221153363 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.466643] env[62000]: DEBUG nova.compute.manager [req-2dd10f33-cc99-452e-a504-4f144ce0c1d0 req-9db6b71f-0d90-4b0f-9100-3db3df70b2a0 service nova] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Detach interface failed, port_id=27537c6a-a039-4b84-877b-e91b648742e0, reason: Instance 23366d62-294b-47c8-adc6-80e854f452ef could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 970.618749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ec4df866-e714-4a9a-b371-1a3968f5c313 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.251s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.731396] env[62000]: INFO nova.compute.manager [-] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Took 1.24 seconds to deallocate network for instance. [ 970.736722] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882710, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.747665] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882704, 'name': CloneVM_Task, 'duration_secs': 1.585775} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.747938] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Created linked-clone VM from snapshot [ 970.748758] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2e81d2f-88bf-4749-a45b-4bd2e3ad44cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.756667] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.782s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.758658] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Uploading image 2877ad1a-5ded-446f-9018-971597893fd7 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 970.760779] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.998s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.762457] env[62000]: INFO nova.compute.claims [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 970.771821] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 970.773200] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-b94afa98-9649-4fd1-9f87-56b443ac3665 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.782482] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 970.782482] env[62000]: value = "task-882711" [ 970.782482] env[62000]: _type = "Task" [ 970.782482] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.786657] env[62000]: INFO nova.scheduler.client.report [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted allocations for instance 94605ab8-8167-430e-b1cd-c8f51e50d8b9 [ 970.793621] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882711, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.821281] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.821545] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.822147] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.822147] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.822329] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.824510] env[62000]: INFO nova.compute.manager [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Terminating instance [ 970.827049] env[62000]: DEBUG nova.compute.manager [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.827257] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 970.828077] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85bb40bf-9ce0-419e-958e-3740516f7e7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.836365] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 970.836606] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3feea711-8a63-42e4-9bd1-4e93ddfb4719 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.842626] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 970.842626] env[62000]: value = "task-882712" [ 970.842626] env[62000]: _type = "Task" [ 970.842626] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.850172] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882712, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.236652] env[62000]: DEBUG oslo_vmware.api [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882710, 'name': PowerOnVM_Task, 'duration_secs': 0.734494} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.237227] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 971.237358] env[62000]: INFO nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Took 9.15 seconds to spawn the instance on the hypervisor. [ 971.237539] env[62000]: DEBUG nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 971.238556] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb151555-eb20-414b-b7a6-1d32bb4959a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.243861] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.293206] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882711, 'name': Destroy_Task, 'duration_secs': 0.501489} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.293486] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Destroyed the VM [ 971.293718] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 971.293966] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ac71aa6a-1e46-4a1c-98b4-99048498f050 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.298432] env[62000]: DEBUG oslo_concurrency.lockutils [None req-26fb10ea-8601-4f3f-a8e4-31d196f1fc6f tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "94605ab8-8167-430e-b1cd-c8f51e50d8b9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.793s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.301363] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 971.301363] env[62000]: value = "task-882713" [ 971.301363] env[62000]: _type = "Task" [ 971.301363] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.309253] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882713, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.352223] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882712, 'name': PowerOffVM_Task, 'duration_secs': 0.219806} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.352492] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 971.352667] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 971.352910] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-76e6485c-defe-4db1-a0be-3b1043cd1223 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.543050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.543293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.756960] env[62000]: INFO nova.compute.manager [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Took 16.67 seconds to build instance. [ 971.813477] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882713, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.957600] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 971.957891] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 971.958131] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleting the datastore file [datastore1] 3780b22d-c360-4433-9f6b-9d5d1f14b525 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.958450] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bcf6d28e-6c48-4e36-a99b-dacf5c2813b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.966470] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 971.966470] env[62000]: value = "task-882716" [ 971.966470] env[62000]: _type = "Task" [ 971.966470] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.979678] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882716, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.981378] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-376d8fea-7fd0-4506-9d0f-87000ffa806c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.988882] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bba6707-3392-45d1-9c29-9705242a98a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.021869] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.022262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.022520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.023008] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.023008] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.025043] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1876a16-a4ca-4a59-83be-1dcc6a8b48de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.028302] env[62000]: INFO nova.compute.manager [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Terminating instance [ 972.030853] env[62000]: DEBUG nova.compute.manager [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.030853] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.031555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7a3e41-f37d-4f2d-82d2-a92d70cfb703 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.040456] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75ae0719-535e-4830-8d8c-f603f5072d7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.046703] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 972.049396] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.050742] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6741129e-3715-4c18-8333-fceab93a38ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.060937] env[62000]: DEBUG nova.compute.provider_tree [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 972.066993] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 972.066993] env[62000]: value = "task-882717" [ 972.066993] env[62000]: _type = "Task" [ 972.066993] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.077185] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.258880] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f76ef10d-bed5-4369-bc6d-0843f06f0dd0 tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.175s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.311986] env[62000]: DEBUG oslo_vmware.api [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882713, 'name': RemoveSnapshot_Task, 'duration_secs': 0.541951} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.312306] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 972.478171] env[62000]: DEBUG oslo_vmware.api [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882716, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266672} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.478416] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.478606] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 972.478791] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 972.478965] env[62000]: INFO nova.compute.manager [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Took 1.65 seconds to destroy the instance on the hypervisor. [ 972.479225] env[62000]: DEBUG oslo.service.loopingcall [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.479420] env[62000]: DEBUG nova.compute.manager [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.479515] env[62000]: DEBUG nova.network.neutron [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 972.563828] env[62000]: DEBUG nova.scheduler.client.report [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.571973] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.577516] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882717, 'name': PowerOffVM_Task, 'duration_secs': 0.179305} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.577773] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 972.577948] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 972.578207] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38086db7-26ba-4914-a9b4-30f77bdd9c7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.635413] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 972.635678] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 972.635834] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleting the datastore file [datastore1] 72a5bf70-dc6e-4887-abb8-8fbad64bb065 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 972.636117] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-72298a3b-6e5c-427b-9c26-195fc6690dad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.642224] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for the task: (returnval){ [ 972.642224] env[62000]: value = "task-882719" [ 972.642224] env[62000]: _type = "Task" [ 972.642224] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.649777] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.818239] env[62000]: WARNING nova.compute.manager [None req-490f1993-d50e-43e0-aaf8-97bec656f217 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Image not found during snapshot: nova.exception.ImageNotFound: Image 2877ad1a-5ded-446f-9018-971597893fd7 could not be found. [ 972.946038] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.946611] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.946611] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.946767] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.946909] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.948777] env[62000]: INFO nova.compute.manager [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Terminating instance [ 972.951781] env[62000]: DEBUG nova.compute.manager [req-875d02fe-863e-402e-93b0-51d90e1cd4bb req-723e99ef-e4d1-45ee-bfb0-12cb72c84e37 service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Received event network-vif-deleted-7b065f0f-9bdd-4fbe-9992-aeebe892c6d0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 972.952268] env[62000]: INFO nova.compute.manager [req-875d02fe-863e-402e-93b0-51d90e1cd4bb req-723e99ef-e4d1-45ee-bfb0-12cb72c84e37 service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Neutron deleted interface 7b065f0f-9bdd-4fbe-9992-aeebe892c6d0; detaching it from the instance and deleting it from the info cache [ 972.952268] env[62000]: DEBUG nova.network.neutron [req-875d02fe-863e-402e-93b0-51d90e1cd4bb req-723e99ef-e4d1-45ee-bfb0-12cb72c84e37 service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 972.953441] env[62000]: DEBUG nova.compute.manager [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.953702] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 972.954808] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a0b39a2-24c3-4796-ac0e-2db364e5b0c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.963465] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 972.963465] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-8f9758f2-9f5b-4f79-b5cb-7e0a11ca8e0f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.970105] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 972.970105] env[62000]: value = "task-882720" [ 972.970105] env[62000]: _type = "Task" [ 972.970105] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.978040] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882720, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.069323] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.069958] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 973.072761] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.829s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.072990] env[62000]: DEBUG nova.objects.instance [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid 23366d62-294b-47c8-adc6-80e854f452ef {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 973.156869] env[62000]: DEBUG oslo_vmware.api [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Task: {'id': task-882719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.367567} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.157104] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.157436] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 973.157802] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 973.158092] env[62000]: INFO nova.compute.manager [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Took 1.13 seconds to destroy the instance on the hypervisor. [ 973.158422] env[62000]: DEBUG oslo.service.loopingcall [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.158703] env[62000]: DEBUG nova.compute.manager [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.158854] env[62000]: DEBUG nova.network.neutron [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 973.431246] env[62000]: DEBUG nova.network.neutron [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.458464] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-54fa8cd2-e085-4a28-8d78-630880785a4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.467291] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-612abb60-b4d9-4834-9007-5e372b70fcc7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.487444] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882720, 'name': PowerOffVM_Task, 'duration_secs': 0.186357} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.487444] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 973.487444] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 973.487444] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b09c0ec5-be23-4217-bba8-a69091a0ced9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.502542] env[62000]: DEBUG nova.compute.manager [req-875d02fe-863e-402e-93b0-51d90e1cd4bb req-723e99ef-e4d1-45ee-bfb0-12cb72c84e37 service nova] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Detach interface failed, port_id=7b065f0f-9bdd-4fbe-9992-aeebe892c6d0, reason: Instance 3780b22d-c360-4433-9f6b-9d5d1f14b525 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 973.573150] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 973.573366] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 973.573581] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleting the datastore file [datastore1] 0e66a948-61d5-4991-99bd-374db7eaaf2a {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.573858] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2005d813-b036-4451-9a67-307fe6348ad5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.576802] env[62000]: DEBUG nova.compute.utils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 973.581244] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 973.581295] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 973.588465] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for the task: (returnval){ [ 973.588465] env[62000]: value = "task-882722" [ 973.588465] env[62000]: _type = "Task" [ 973.588465] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.597186] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882722, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.621803] env[62000]: DEBUG nova.policy [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 973.684950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.685282] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.685504] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.685694] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.685870] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.688013] env[62000]: INFO nova.compute.manager [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Terminating instance [ 973.690068] env[62000]: DEBUG nova.compute.manager [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 973.690272] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 973.691239] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f79f26-c12d-4037-b315-f2c51d4023f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.702575] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 973.702876] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aab538fa-e923-418f-aec2-7390e2331b5a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.710648] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 973.710648] env[62000]: value = "task-882723" [ 973.710648] env[62000]: _type = "Task" [ 973.710648] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.718654] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882723, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.802799] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7815f521-c8b2-47f2-8e4c-0736ca01c07b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.811115] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 973.811115] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201628', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'name': 'volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3878579f-6435-4fe3-9f8c-8461d8ac57ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'serial': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 973.811115] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6024ae0-89f0-4ed8-9b1e-9bf3dad29ada {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.815021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2784fc0d-cbe5-4222-89ba-467089b3d637 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.830087] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e5283f-f497-4b22-a32f-e0e84b907850 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.861017] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31828e6-ed17-4522-926c-4ab27069e58d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.883744] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7/volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 973.884529] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-105d15be-889e-453a-a9a8-d4ff51a0bf65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.901143] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c96d6d8-f81c-4014-91a6-f5d031ec0d12 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.908296] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 973.908296] env[62000]: value = "task-882724" [ 973.908296] env[62000]: _type = "Task" [ 973.908296] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.921015] env[62000]: DEBUG nova.compute.provider_tree [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.925365] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882724, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.934357] env[62000]: INFO nova.compute.manager [-] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Took 1.45 seconds to deallocate network for instance. [ 973.959094] env[62000]: DEBUG nova.network.neutron [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.997354] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Successfully created port: b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 974.081873] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 974.100057] env[62000]: DEBUG oslo_vmware.api [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Task: {'id': task-882722, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135639} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.100057] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.100057] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 974.100057] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 974.100057] env[62000]: INFO nova.compute.manager [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Took 1.15 seconds to destroy the instance on the hypervisor. [ 974.100495] env[62000]: DEBUG oslo.service.loopingcall [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.100495] env[62000]: DEBUG nova.compute.manager [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 974.100606] env[62000]: DEBUG nova.network.neutron [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 974.221916] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882723, 'name': PowerOffVM_Task, 'duration_secs': 0.206484} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.222366] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 974.222633] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 974.222912] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86e2d073-b44e-4ef0-bd70-a3fe1017d28e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.294984] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 974.295271] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 974.295466] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleting the datastore file [datastore2] b6579b75-42dc-473d-8b66-24aa39f9e4ff {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 974.295739] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21134255-2f20-486e-a1cf-34bb5f9430e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.304502] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for the task: (returnval){ [ 974.304502] env[62000]: value = "task-882726" [ 974.304502] env[62000]: _type = "Task" [ 974.304502] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.313851] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882726, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.419835] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882724, 'name': ReconfigVM_Task, 'duration_secs': 0.416856} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.419835] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to attach disk [datastore2] volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7/volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.425740] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0517e533-d705-49f9-86f9-be600119cf36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.437949] env[62000]: DEBUG nova.scheduler.client.report [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 974.441200] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.447012] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 974.447012] env[62000]: value = "task-882727" [ 974.447012] env[62000]: _type = "Task" [ 974.447012] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.456830] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882727, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.461584] env[62000]: INFO nova.compute.manager [-] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Took 1.30 seconds to deallocate network for instance. [ 974.815353] env[62000]: DEBUG oslo_vmware.api [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Task: {'id': task-882726, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141872} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.815638] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 974.815830] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 974.816021] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 974.816205] env[62000]: INFO nova.compute.manager [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Took 1.13 seconds to destroy the instance on the hypervisor. [ 974.816447] env[62000]: DEBUG oslo.service.loopingcall [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 974.816645] env[62000]: DEBUG nova.compute.manager [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 974.816743] env[62000]: DEBUG nova.network.neutron [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 974.936769] env[62000]: DEBUG nova.network.neutron [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.943463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.871s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.945806] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.374s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.947576] env[62000]: INFO nova.compute.claims [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 974.960675] env[62000]: DEBUG oslo_vmware.api [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882727, 'name': ReconfigVM_Task, 'duration_secs': 0.141448} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.961186] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201628', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'name': 'volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3878579f-6435-4fe3-9f8c-8461d8ac57ee', 'attached_at': '', 'detached_at': '', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'serial': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 974.967638] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.978452] env[62000]: INFO nova.scheduler.client.report [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance 23366d62-294b-47c8-adc6-80e854f452ef [ 974.992126] env[62000]: DEBUG nova.compute.manager [req-ed62f98a-7612-4f9f-8f1b-56c014ef4e7e req-208ff0f5-4eb4-499f-a8bd-b4175bb752a9 service nova] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Received event network-vif-deleted-f0d1137e-2373-4645-85db-8c5c8c963ff4 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 974.992349] env[62000]: DEBUG nova.compute.manager [req-ed62f98a-7612-4f9f-8f1b-56c014ef4e7e req-208ff0f5-4eb4-499f-a8bd-b4175bb752a9 service nova] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Received event network-vif-deleted-a7dcf91b-1563-4262-86e8-ee28ac8d0bbe {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 975.024686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.025040] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.025523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "66edb906-f936-4f57-833d-224f36af109e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.025954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.026035] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.028243] env[62000]: INFO nova.compute.manager [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Terminating instance [ 975.031379] env[62000]: DEBUG nova.compute.manager [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 975.031579] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 975.032475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51997d9b-15d9-45d0-b033-f7fb364b8806 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.041078] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 975.041321] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-71ea84ce-83b9-46a4-ac0b-5d982eaaa9e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.049108] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 975.049108] env[62000]: value = "task-882728" [ 975.049108] env[62000]: _type = "Task" [ 975.049108] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.058707] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882728, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.090713] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 975.118229] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 975.118538] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 975.118724] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 975.118927] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 975.119117] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 975.119293] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 975.119525] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 975.119713] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 975.119898] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 975.120112] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 975.120326] env[62000]: DEBUG nova.virt.hardware [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 975.121379] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c55068e6-a459-4635-b1d7-048b7cfe84b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.129466] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e962de-a14d-45b6-99d8-ed2f14932543 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.440248] env[62000]: INFO nova.compute.manager [-] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Took 1.34 seconds to deallocate network for instance. [ 975.497073] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0ec2844f-5e5c-4eb9-a39b-5af0b4b1ace5 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "23366d62-294b-47c8-adc6-80e854f452ef" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.119s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.561217] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882728, 'name': PowerOffVM_Task, 'duration_secs': 0.180395} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.561530] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 975.561707] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 975.561967] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-50ae8362-4d45-4f99-b7da-7cab934cf385 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.570816] env[62000]: DEBUG nova.network.neutron [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 975.616951] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Successfully updated port: b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 975.946964] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.963362] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 975.963626] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 975.963814] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleting the datastore file [datastore2] 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 975.964362] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4da2225e-be51-4411-a712-17a0f824fa65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.970846] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 975.970846] env[62000]: value = "task-882730" [ 975.970846] env[62000]: _type = "Task" [ 975.970846] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.978789] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882730, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.021092] env[62000]: DEBUG nova.objects.instance [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'flavor' on Instance uuid 3878579f-6435-4fe3-9f8c-8461d8ac57ee {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.049618] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.049847] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.050081] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.050291] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.050465] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.052519] env[62000]: INFO nova.compute.manager [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Terminating instance [ 976.054397] env[62000]: DEBUG nova.compute.manager [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 976.054596] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 976.055420] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf820bfd-a3d3-447d-9a2d-34840a185f9e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.065871] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 976.066192] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-867d198c-e099-4c53-95d7-a5013afa6c86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.072845] env[62000]: INFO nova.compute.manager [-] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Took 1.26 seconds to deallocate network for instance. [ 976.073130] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 976.073130] env[62000]: value = "task-882731" [ 976.073130] env[62000]: _type = "Task" [ 976.073130] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.089501] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.121810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 976.121963] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 976.122163] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 976.158400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cc06eb7-2909-45c5-982f-394fcf8f0f9e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.166972] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f38102-a051-4750-8515-d9d487d24e7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.203912] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa721b0a-bce3-4977-9228-c07763d31994 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.213179] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e053417-35fb-429f-898f-98de809b922a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.227948] env[62000]: DEBUG nova.compute.provider_tree [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 976.563062] env[62000]: DEBUG oslo_vmware.api [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882730, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.163182} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.563062] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 976.563062] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 976.563062] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 976.563062] env[62000]: INFO nova.compute.manager [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 66edb906-f936-4f57-833d-224f36af109e] Took 1.45 seconds to destroy the instance on the hypervisor. [ 976.563062] env[62000]: DEBUG oslo.service.loopingcall [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 976.563062] env[62000]: DEBUG nova.compute.manager [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 976.563062] env[62000]: DEBUG nova.network.neutron [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 976.563062] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29a08717-e4f6-4c4b-8044-89781f6560a5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.332s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.587944] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 976.588273] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882731, 'name': PowerOffVM_Task, 'duration_secs': 0.211557} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.588525] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 976.588697] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 976.588943] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d0a0980-bdb4-4b68-8eea-8b500d3c1f53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.652017] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 976.652342] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 976.652562] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore2] ea58da8c-e4ac-4863-942d-6294cf04e6f2 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 976.652837] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-65dedccd-057d-4c22-976f-5fbb218b934c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.659355] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 976.659355] env[62000]: value = "task-882733" [ 976.659355] env[62000]: _type = "Task" [ 976.659355] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.667082] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882733, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.680723] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 976.733580] env[62000]: DEBUG nova.scheduler.client.report [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.936894] env[62000]: DEBUG nova.network.neutron [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Updating instance_info_cache with network_info: [{"id": "b790e04b-9158-4ad2-9e49-34e012d38db5", "address": "fa:16:3e:44:66:67", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb790e04b-91", "ovs_interfaceid": "b790e04b-9158-4ad2-9e49-34e012d38db5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.028655] env[62000]: DEBUG nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Received event network-vif-deleted-fb11497c-0f1d-4ff6-baf0-970f2e80bc28 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.029018] env[62000]: DEBUG nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Received event network-vif-plugged-b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.029262] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Acquiring lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.029514] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.029686] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.029856] env[62000]: DEBUG nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] No waiting events found dispatching network-vif-plugged-b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.030037] env[62000]: WARNING nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Received unexpected event network-vif-plugged-b790e04b-9158-4ad2-9e49-34e012d38db5 for instance with vm_state building and task_state spawning. [ 977.030250] env[62000]: DEBUG nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Received event network-changed-b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 977.030464] env[62000]: DEBUG nova.compute.manager [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Refreshing instance network info cache due to event network-changed-b790e04b-9158-4ad2-9e49-34e012d38db5. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 977.030525] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Acquiring lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.173109] env[62000]: DEBUG oslo_vmware.api [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882733, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134555} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.173227] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 977.173496] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 977.173763] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 977.174325] env[62000]: INFO nova.compute.manager [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Took 1.12 seconds to destroy the instance on the hypervisor. [ 977.174425] env[62000]: DEBUG oslo.service.loopingcall [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.174737] env[62000]: DEBUG nova.compute.manager [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 977.174824] env[62000]: DEBUG nova.network.neutron [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 977.243584] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.298s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.244167] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 977.246832] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.806s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.247014] env[62000]: DEBUG nova.objects.instance [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'resources' on Instance uuid 3780b22d-c360-4433-9f6b-9d5d1f14b525 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.350384] env[62000]: DEBUG nova.compute.manager [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 977.440148] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.440460] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Instance network_info: |[{"id": "b790e04b-9158-4ad2-9e49-34e012d38db5", "address": "fa:16:3e:44:66:67", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb790e04b-91", "ovs_interfaceid": "b790e04b-9158-4ad2-9e49-34e012d38db5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 977.440813] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Acquired lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.441018] env[62000]: DEBUG nova.network.neutron [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Refreshing network info cache for port b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 977.445172] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:44:66:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b790e04b-9158-4ad2-9e49-34e012d38db5', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 977.450229] env[62000]: DEBUG oslo.service.loopingcall [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 977.451161] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 977.451420] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9ead464d-a0fb-47d9-871a-7e772129e80a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.472238] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 977.472238] env[62000]: value = "task-882734" [ 977.472238] env[62000]: _type = "Task" [ 977.472238] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.479987] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882734, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.609654] env[62000]: DEBUG nova.network.neutron [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.750630] env[62000]: DEBUG nova.compute.utils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 977.753079] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 977.753290] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 977.796521] env[62000]: DEBUG nova.policy [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79730e91610c4c598019b122b4f41418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6106a70abdad4111977eebbd82434337', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 977.869313] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.930706] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329c3e0d-834c-4e52-b164-7896d7bb7738 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.937667] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b430e74d-cc49-42fd-ac01-5a1d74ccc877 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.974059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe31c5b-3bb0-417a-8fe3-c84f1212db31 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.981819] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882734, 'name': CreateVM_Task, 'duration_secs': 0.492366} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.983679] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 977.984399] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.984561] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.984884] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 977.986163] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010d95d6-c349-410e-b704-f5a00e3949c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.989677] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b45169d1-3ca8-4be4-8e84-a7af476fa515 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.994643] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 977.994643] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fb0e02-9d77-5140-f15e-524259b87376" [ 977.994643] env[62000]: _type = "Task" [ 977.994643] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.002344] env[62000]: DEBUG nova.compute.provider_tree [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 978.011056] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fb0e02-9d77-5140-f15e-524259b87376, 'name': SearchDatastore_Task, 'duration_secs': 0.009544} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.011562] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.011863] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 978.012149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 978.012312] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 978.012513] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 978.013287] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d058e8a2-a285-4f2a-be5f-0cb69b929e75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.018076] env[62000]: DEBUG nova.network.neutron [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.020550] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 978.020733] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 978.023366] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf61c3b4-29f5-404d-a693-dcb19e518d06 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.028328] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 978.028328] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526db2d8-277e-9a33-be32-83878c4147d2" [ 978.028328] env[62000]: _type = "Task" [ 978.028328] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.036535] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526db2d8-277e-9a33-be32-83878c4147d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.112092] env[62000]: INFO nova.compute.manager [-] [instance: 66edb906-f936-4f57-833d-224f36af109e] Took 1.63 seconds to deallocate network for instance. [ 978.179735] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Successfully created port: 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 978.253973] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 978.301825] env[62000]: DEBUG nova.network.neutron [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Updated VIF entry in instance network info cache for port b790e04b-9158-4ad2-9e49-34e012d38db5. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 978.302297] env[62000]: DEBUG nova.network.neutron [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Updating instance_info_cache with network_info: [{"id": "b790e04b-9158-4ad2-9e49-34e012d38db5", "address": "fa:16:3e:44:66:67", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb790e04b-91", "ovs_interfaceid": "b790e04b-9158-4ad2-9e49-34e012d38db5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.508460] env[62000]: DEBUG nova.scheduler.client.report [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.520339] env[62000]: INFO nova.compute.manager [-] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Took 1.35 seconds to deallocate network for instance. [ 978.539311] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526db2d8-277e-9a33-be32-83878c4147d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009442} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.540295] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf9200d0-d968-4ccd-8d6e-37846313d52b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.545670] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 978.545670] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523acdbc-5dc6-7ad2-6a61-3ff4659a7408" [ 978.545670] env[62000]: _type = "Task" [ 978.545670] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.553986] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523acdbc-5dc6-7ad2-6a61-3ff4659a7408, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.619712] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.805327] env[62000]: DEBUG oslo_concurrency.lockutils [req-34d244e8-55f9-489d-9d9b-74a698ae47b8 req-b3e83d29-d10b-4797-96f1-742a8bbb1745 service nova] Releasing lock "refresh_cache-c9758d78-83a0-430e-a69e-e2d82f99daa9" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.013284] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.766s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.015863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.048s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.016347] env[62000]: DEBUG nova.objects.instance [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lazy-loading 'resources' on Instance uuid 72a5bf70-dc6e-4887-abb8-8fbad64bb065 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.026896] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.036747] env[62000]: INFO nova.scheduler.client.report [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted allocations for instance 3780b22d-c360-4433-9f6b-9d5d1f14b525 [ 979.056250] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523acdbc-5dc6-7ad2-6a61-3ff4659a7408, 'name': SearchDatastore_Task, 'duration_secs': 0.010051} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.056549] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.056815] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c9758d78-83a0-430e-a69e-e2d82f99daa9/c9758d78-83a0-430e-a69e-e2d82f99daa9.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 979.057379] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ca6d5503-cbd5-4921-9faf-f9f46349f390 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.065929] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 979.065929] env[62000]: value = "task-882735" [ 979.065929] env[62000]: _type = "Task" [ 979.065929] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.074461] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882735, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.088777] env[62000]: DEBUG nova.compute.manager [req-fb688e64-0a52-4b18-98bc-a1b4bf6fb5f1 req-723bff75-bb05-46ee-84b7-048e3d39d595 service nova] [instance: 66edb906-f936-4f57-833d-224f36af109e] Received event network-vif-deleted-849930d2-4200-4ec9-ac15-51c243a118bf {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.089019] env[62000]: DEBUG nova.compute.manager [req-fb688e64-0a52-4b18-98bc-a1b4bf6fb5f1 req-723bff75-bb05-46ee-84b7-048e3d39d595 service nova] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Received event network-vif-deleted-86034d5a-0fb2-425b-a306-aa065f14015a {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 979.263683] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 979.292633] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 979.292902] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 979.293083] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 979.293324] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 979.293483] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 979.293658] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 979.293845] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 979.294014] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 979.294194] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 979.294386] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 979.294556] env[62000]: DEBUG nova.virt.hardware [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 979.295462] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1b5902-1b77-484b-94aa-8ca35e7cb9a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.303446] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6df8de62-1140-4b5e-992b-62d0d3b22898 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.549817] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bbb2d73f-cdfd-441c-9d6c-4da2c79248ce tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "3780b22d-c360-4433-9f6b-9d5d1f14b525" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.725s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 979.580211] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882735, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444007} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.580607] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c9758d78-83a0-430e-a69e-e2d82f99daa9/c9758d78-83a0-430e-a69e-e2d82f99daa9.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 979.580779] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 979.580937] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4d3fbab-877b-4c36-97e2-937d01f01e92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.587582] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 979.587582] env[62000]: value = "task-882736" [ 979.587582] env[62000]: _type = "Task" [ 979.587582] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.601407] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882736, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.706475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0113e0f-216e-4f08-88e9-71becf5c8cc4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.715318] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4406cf53-4f47-452d-ac56-c8f16fa25a17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.750741] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4364737b-3ed1-4b24-ad57-b86d82cec1d2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.758600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0197524-f392-4508-a9d9-1d1214604dda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.773352] env[62000]: DEBUG nova.compute.provider_tree [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.958702] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Successfully updated port: 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 980.097011] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882736, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062169} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.097424] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 980.098228] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d9ed94-c451-423c-acad-8f64e0794322 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.127327] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] c9758d78-83a0-430e-a69e-e2d82f99daa9/c9758d78-83a0-430e-a69e-e2d82f99daa9.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 980.129452] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77d86296-c0cc-4aef-8f33-0444467f9e11 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.150458] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 980.150458] env[62000]: value = "task-882737" [ 980.150458] env[62000]: _type = "Task" [ 980.150458] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.161446] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882737, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.278993] env[62000]: DEBUG nova.scheduler.client.report [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.463042] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 980.463298] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 980.463442] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 980.660232] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882737, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.784158] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.786484] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.840s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.786727] env[62000]: DEBUG nova.objects.instance [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lazy-loading 'resources' on Instance uuid 0e66a948-61d5-4991-99bd-374db7eaaf2a {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 980.809160] env[62000]: INFO nova.scheduler.client.report [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Deleted allocations for instance 72a5bf70-dc6e-4887-abb8-8fbad64bb065 [ 981.003545] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 981.115878] env[62000]: DEBUG nova.compute.manager [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Received event network-vif-plugged-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.116738] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Acquiring lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.117112] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.117436] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.117912] env[62000]: DEBUG nova.compute.manager [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] No waiting events found dispatching network-vif-plugged-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.118293] env[62000]: WARNING nova.compute.manager [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Received unexpected event network-vif-plugged-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb for instance with vm_state building and task_state spawning. [ 981.118546] env[62000]: DEBUG nova.compute.manager [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Received event network-changed-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 981.118848] env[62000]: DEBUG nova.compute.manager [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Refreshing instance network info cache due to event network-changed-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 981.119172] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Acquiring lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.164810] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882737, 'name': ReconfigVM_Task, 'duration_secs': 0.527186} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.165142] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Reconfigured VM instance instance-0000005d to attach disk [datastore1] c9758d78-83a0-430e-a69e-e2d82f99daa9/c9758d78-83a0-430e-a69e-e2d82f99daa9.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 981.165942] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f6044ab3-959f-4084-bb97-66830e2480b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.173749] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 981.173749] env[62000]: value = "task-882738" [ 981.173749] env[62000]: _type = "Task" [ 981.173749] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.193495] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882738, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.257277] env[62000]: DEBUG nova.network.neutron [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updating instance_info_cache with network_info: [{"id": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "address": "fa:16:3e:85:86:31", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a0a660-2c", "ovs_interfaceid": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.319018] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d0ca7ee9-3a15-4c96-bae8-532a5b683169 tempest-AttachInterfacesTestJSON-1020067435 tempest-AttachInterfacesTestJSON-1020067435-project-member] Lock "72a5bf70-dc6e-4887-abb8-8fbad64bb065" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.297s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.443373] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0661bf-4a89-49cc-8d71-c472ce0a454b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.451340] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5bb13e9-9754-4266-8097-27cb0cce18a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.481342] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ae10520-45bf-467b-984e-455f727b0de3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.488244] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.488477] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.492750] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b145df-e3bd-4166-accd-dad8e26c7fc9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.506439] env[62000]: DEBUG nova.compute.provider_tree [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.685931] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882738, 'name': Rename_Task, 'duration_secs': 0.133507} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.686418] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 981.686797] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2b69cc6d-4687-42e4-8e31-6e90b7d3be4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.693824] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 981.693824] env[62000]: value = "task-882739" [ 981.693824] env[62000]: _type = "Task" [ 981.693824] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.700958] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882739, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.761533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.762046] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Instance network_info: |[{"id": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "address": "fa:16:3e:85:86:31", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a0a660-2c", "ovs_interfaceid": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 981.762523] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Acquired lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.762813] env[62000]: DEBUG nova.network.neutron [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Refreshing network info cache for port 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 981.764432] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:85:86:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40a0a660-2c5f-4d14-829e-1c85ff3fc2bb', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 981.776126] env[62000]: DEBUG oslo.service.loopingcall [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 981.777155] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 981.777395] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-98e8581a-d417-406d-be0b-386a262b4d6f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.797348] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 981.797348] env[62000]: value = "task-882740" [ 981.797348] env[62000]: _type = "Task" [ 981.797348] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.805336] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882740, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.992747] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 982.009845] env[62000]: DEBUG nova.scheduler.client.report [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.204166] env[62000]: DEBUG oslo_vmware.api [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882739, 'name': PowerOnVM_Task, 'duration_secs': 0.437545} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.204500] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 982.204711] env[62000]: INFO nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Took 7.11 seconds to spawn the instance on the hypervisor. [ 982.204898] env[62000]: DEBUG nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.207324] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142fe74a-b0f9-4a7c-8f64-0e21f4c5434e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.309879] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882740, 'name': CreateVM_Task, 'duration_secs': 0.325176} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.310120] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 982.310812] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.310986] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.311834] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 982.311834] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2106a2ee-ddfc-4e91-b5f1-6a5c5d184b1f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.316261] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 982.316261] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523b754d-e830-9fd8-b778-46591aeee802" [ 982.316261] env[62000]: _type = "Task" [ 982.316261] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.323961] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523b754d-e830-9fd8-b778-46591aeee802, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.515706] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.516580] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.730s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.519413] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.931s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.519413] env[62000]: DEBUG nova.objects.instance [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lazy-loading 'resources' on Instance uuid b6579b75-42dc-473d-8b66-24aa39f9e4ff {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 982.537338] env[62000]: INFO nova.scheduler.client.report [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Deleted allocations for instance 0e66a948-61d5-4991-99bd-374db7eaaf2a [ 982.583458] env[62000]: DEBUG nova.network.neutron [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updated VIF entry in instance network info cache for port 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 982.583839] env[62000]: DEBUG nova.network.neutron [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updating instance_info_cache with network_info: [{"id": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "address": "fa:16:3e:85:86:31", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a0a660-2c", "ovs_interfaceid": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.721419] env[62000]: INFO nova.compute.manager [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Took 14.97 seconds to build instance. [ 982.815243] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.815477] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 982.827635] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523b754d-e830-9fd8-b778-46591aeee802, 'name': SearchDatastore_Task, 'duration_secs': 0.008782} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.827845] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.828091] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 982.828330] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.831078] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.831078] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 982.831078] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dd5b9445-bbe4-48f1-bb35-b42550291a22 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.838331] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 982.838522] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 982.839505] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bf83acc-c3ac-411e-80d9-682414cd0bca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.845368] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 982.845368] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5204ae9b-6ab1-12d7-8772-eff1a1d4b9f3" [ 982.845368] env[62000]: _type = "Task" [ 982.845368] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.853393] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5204ae9b-6ab1-12d7-8772-eff1a1d4b9f3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.044044] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d719308b-c0cb-4cca-9006-0bc61946bb5f tempest-ServerDiskConfigTestJSON-1850537075 tempest-ServerDiskConfigTestJSON-1850537075-project-member] Lock "0e66a948-61d5-4991-99bd-374db7eaaf2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.097s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.087364] env[62000]: DEBUG oslo_concurrency.lockutils [req-07869e1f-9327-497d-bf5d-e5edd1bc9bfe req-f2dc87ce-e975-44ef-97c4-a3df1856d4fb service nova] Releasing lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.218576] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34272792-60d1-4309-8903-db5481175af8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.223583] env[62000]: DEBUG oslo_concurrency.lockutils [None req-bd846c17-3813-4a1b-85d4-5ec00b3f32a8 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.485s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.226751] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba58ebe1-f48b-42e3-be82-50a0d764cdd2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.259584] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e02216b-ed6a-4780-881c-02a3d7d613ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.268477] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ca2dd6-b35f-4afb-99db-fa05740e6b9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.281613] env[62000]: DEBUG nova.compute.provider_tree [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 983.324948] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 983.324948] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 983.325127] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Rebuilding the list of instances to heal {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 983.358554] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5204ae9b-6ab1-12d7-8772-eff1a1d4b9f3, 'name': SearchDatastore_Task, 'duration_secs': 0.010292} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.359593] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc008480-5a9b-425a-8e42-967f373ceea0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.364725] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 983.364725] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52995b1a-8f70-492c-3936-d456c43fa93e" [ 983.364725] env[62000]: _type = "Task" [ 983.364725] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.372366] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52995b1a-8f70-492c-3936-d456c43fa93e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.595188] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.595495] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.595694] env[62000]: DEBUG nova.compute.manager [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 983.596625] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43cf3841-fd04-4739-a6a5-e3dfbdfecb4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.603355] env[62000]: DEBUG nova.compute.manager [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 983.603930] env[62000]: DEBUG nova.objects.instance [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'flavor' on Instance uuid c9758d78-83a0-430e-a69e-e2d82f99daa9 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.784871] env[62000]: DEBUG nova.scheduler.client.report [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 983.833628] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Skipping network cache update for instance because it is Building. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 983.869909] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.870075] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquired lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 983.870223] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Forcefully refreshing network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 983.870378] env[62000]: DEBUG nova.objects.instance [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lazy-loading 'info_cache' on Instance uuid f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 983.877963] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52995b1a-8f70-492c-3936-d456c43fa93e, 'name': SearchDatastore_Task, 'duration_secs': 0.008594} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.878310] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.878585] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f8f26be0-f7d1-43e9-886d-c19d385e6935/f8f26be0-f7d1-43e9-886d-c19d385e6935.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 983.878856] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38b9dfa7-35a4-4423-9672-dc74ae6b91f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.885764] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 983.885764] env[62000]: value = "task-882741" [ 983.885764] env[62000]: _type = "Task" [ 983.885764] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.894959] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882741, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.111484] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 984.111484] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e6c3e1e-b27b-4df0-b6ac-da7d11494f83 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.119215] env[62000]: DEBUG oslo_vmware.api [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 984.119215] env[62000]: value = "task-882742" [ 984.119215] env[62000]: _type = "Task" [ 984.119215] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.127504] env[62000]: DEBUG oslo_vmware.api [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882742, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.290190] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.771s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.294112] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.424s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 984.323955] env[62000]: INFO nova.scheduler.client.report [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Deleted allocations for instance b6579b75-42dc-473d-8b66-24aa39f9e4ff [ 984.399027] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882741, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.502938} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.399618] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] f8f26be0-f7d1-43e9-886d-c19d385e6935/f8f26be0-f7d1-43e9-886d-c19d385e6935.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 984.399771] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 984.400586] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-92c6680d-f674-453b-8575-9e20f41e84a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.408096] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 984.408096] env[62000]: value = "task-882743" [ 984.408096] env[62000]: _type = "Task" [ 984.408096] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.422681] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882743, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.630978] env[62000]: DEBUG oslo_vmware.api [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882742, 'name': PowerOffVM_Task, 'duration_secs': 0.245197} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.631308] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 984.631515] env[62000]: DEBUG nova.compute.manager [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.632402] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d726a72f-8bf2-4ae2-9aae-c18f0f215a92 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.798190] env[62000]: INFO nova.compute.claims [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 984.835730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3c9528b9-9c96-40c6-8491-b9b32e2053f7 tempest-ImagesTestJSON-1919675188 tempest-ImagesTestJSON-1919675188-project-member] Lock "b6579b75-42dc-473d-8b66-24aa39f9e4ff" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.150s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.919961] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882743, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072132} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.920410] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 984.921305] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad0ecbc-2b3f-40b2-a8c7-388ef71911dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.945582] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] f8f26be0-f7d1-43e9-886d-c19d385e6935/f8f26be0-f7d1-43e9-886d-c19d385e6935.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.947525] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f52b0869-c282-4c17-b87c-e4b55d0321f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.972951] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 984.972951] env[62000]: value = "task-882744" [ 984.972951] env[62000]: _type = "Task" [ 984.972951] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.981073] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882744, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.144583] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f677d8e2-d413-40de-bd31-edbf7ad48d82 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.303275] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.306134] env[62000]: INFO nova.compute.resource_tracker [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating resource usage from migration 5793fe93-6df4-44ae-af50-d526a12184f6 [ 985.487732] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882744, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.490304] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ec78ce5-7e35-47b6-946f-afc9b99a4c30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.501952] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d34a79d-db44-40fb-8f27-b2e70bacebaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.550357] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65f16f3b-8437-4193-b84d-f45b802e8506 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.558683] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be975a8-7f13-4de5-a5f3-795b71789d6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.575031] env[62000]: DEBUG nova.compute.provider_tree [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 985.809485] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Releasing lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.809718] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updated the network info_cache for instance {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 985.809959] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810162] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810317] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810471] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810609] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810783] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.810936] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 985.811108] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 985.988149] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882744, 'name': ReconfigVM_Task, 'duration_secs': 0.98334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.988432] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Reconfigured VM instance instance-0000005e to attach disk [datastore1] f8f26be0-f7d1-43e9-886d-c19d385e6935/f8f26be0-f7d1-43e9-886d-c19d385e6935.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.989197] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8c980242-3457-433d-b918-7375de6826b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.996247] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 985.996247] env[62000]: value = "task-882745" [ 985.996247] env[62000]: _type = "Task" [ 985.996247] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.008429] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882745, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.079893] env[62000]: DEBUG nova.scheduler.client.report [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.315031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.487533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.487533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.487533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.487533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.487533] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.491437] env[62000]: INFO nova.compute.manager [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Terminating instance [ 986.495096] env[62000]: DEBUG nova.compute.manager [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 986.495444] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 986.496688] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c93fe1-104e-4d89-84e2-babbb3e4fdac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.511252] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 986.515044] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d84f0504-070f-44e9-9217-2e784c7dc0e7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.517499] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882745, 'name': Rename_Task, 'duration_secs': 0.159343} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.518238] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 986.519524] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c65a7aa6-8861-4d3a-81ce-056fd7c3d51d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.527232] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 986.527232] env[62000]: value = "task-882747" [ 986.527232] env[62000]: _type = "Task" [ 986.527232] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.536171] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882747, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.586651] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.294s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 986.586895] env[62000]: INFO nova.compute.manager [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Migrating [ 986.595751] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.976s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.595751] env[62000]: DEBUG nova.objects.instance [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'resources' on Instance uuid 66edb906-f936-4f57-833d-224f36af109e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 986.613254] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 986.613254] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 986.613254] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore1] c9758d78-83a0-430e-a69e-e2d82f99daa9 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 986.617619] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-11e7c24a-04fe-4362-8bfd-96f6520b2efb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.627391] env[62000]: DEBUG oslo_vmware.api [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 986.627391] env[62000]: value = "task-882748" [ 986.627391] env[62000]: _type = "Task" [ 986.627391] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.643920] env[62000]: DEBUG oslo_vmware.api [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.045996] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882747, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.113842] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.118022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.118022] env[62000]: DEBUG nova.network.neutron [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 987.139664] env[62000]: DEBUG oslo_vmware.api [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.211751} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.139955] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 987.140235] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 987.140455] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 987.140657] env[62000]: INFO nova.compute.manager [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Took 0.65 seconds to destroy the instance on the hypervisor. [ 987.141279] env[62000]: DEBUG oslo.service.loopingcall [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.141428] env[62000]: DEBUG nova.compute.manager [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 987.141526] env[62000]: DEBUG nova.network.neutron [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 987.312231] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e6f821-2dfd-4d92-ae0d-859a22310cd8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.320136] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740c0ad0-c0a0-4754-bb60-53e18d0265f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.353785] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-416ca890-a0bc-4801-87b1-09267082dd79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.362228] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0524f5d3-e67b-400e-a7ad-84ac13bc3c90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.375466] env[62000]: DEBUG nova.compute.provider_tree [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 987.542238] env[62000]: DEBUG oslo_vmware.api [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882747, 'name': PowerOnVM_Task, 'duration_secs': 0.521493} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.542238] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 987.542238] env[62000]: INFO nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Took 8.28 seconds to spawn the instance on the hypervisor. [ 987.542238] env[62000]: DEBUG nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 987.542429] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94bc0b35-c37b-4edb-87ac-dfa64471070f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.583491] env[62000]: DEBUG nova.compute.manager [req-9bea42d7-21b5-4a6e-a995-abef935cf99f req-f58dfc83-35f1-45d2-a35c-d65685b56b06 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Received event network-vif-deleted-b790e04b-9158-4ad2-9e49-34e012d38db5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 987.583697] env[62000]: INFO nova.compute.manager [req-9bea42d7-21b5-4a6e-a995-abef935cf99f req-f58dfc83-35f1-45d2-a35c-d65685b56b06 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Neutron deleted interface b790e04b-9158-4ad2-9e49-34e012d38db5; detaching it from the instance and deleting it from the info cache [ 987.585510] env[62000]: DEBUG nova.network.neutron [req-9bea42d7-21b5-4a6e-a995-abef935cf99f req-f58dfc83-35f1-45d2-a35c-d65685b56b06 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 987.880069] env[62000]: DEBUG nova.scheduler.client.report [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 987.929778] env[62000]: DEBUG nova.network.neutron [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.048586] env[62000]: DEBUG nova.network.neutron [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.066175] env[62000]: INFO nova.compute.manager [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Took 15.51 seconds to build instance. [ 988.086598] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2831e5f2-d264-4dac-9449-ea3358fecf64 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.096978] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075ceba8-1cf0-4aaf-961f-05cf25e3c372 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.127517] env[62000]: DEBUG nova.compute.manager [req-9bea42d7-21b5-4a6e-a995-abef935cf99f req-f58dfc83-35f1-45d2-a35c-d65685b56b06 service nova] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Detach interface failed, port_id=b790e04b-9158-4ad2-9e49-34e012d38db5, reason: Instance c9758d78-83a0-430e-a69e-e2d82f99daa9 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 988.387077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.391134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.364s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.391655] env[62000]: DEBUG nova.objects.instance [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid ea58da8c-e4ac-4863-942d-6294cf04e6f2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.413326] env[62000]: INFO nova.scheduler.client.report [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleted allocations for instance 66edb906-f936-4f57-833d-224f36af109e [ 988.433206] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.466898] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.467164] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.551077] env[62000]: INFO nova.compute.manager [-] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Took 1.41 seconds to deallocate network for instance. [ 988.570917] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2db81d55-1b87-400c-be3c-5e1b21b6af8d tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.027s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.926614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-0c154912-323e-4fe6-ae09-be20d1daf835 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "66edb906-f936-4f57-833d-224f36af109e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.900s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.971416] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 989.058937] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.093598] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d818c6b-f585-4aaa-852c-d42f4eb55088 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.103488] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-429f29e6-08a6-4a6c-aaf0-a6c31d167c7d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.135854] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5a325b-a7e1-49d7-917b-c96a8c9b5ce0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.144655] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d829ba8-4cc9-45c8-ad3b-2ba8c982a0a3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.160668] env[62000]: DEBUG nova.compute.provider_tree [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 989.269686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.269686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.498895] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.632564] env[62000]: DEBUG nova.compute.manager [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Received event network-changed-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 989.632564] env[62000]: DEBUG nova.compute.manager [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Refreshing instance network info cache due to event network-changed-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 989.632564] env[62000]: DEBUG oslo_concurrency.lockutils [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] Acquiring lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.632564] env[62000]: DEBUG oslo_concurrency.lockutils [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] Acquired lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.633705] env[62000]: DEBUG nova.network.neutron [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Refreshing network info cache for port 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 989.666030] env[62000]: DEBUG nova.scheduler.client.report [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.776130] env[62000]: INFO nova.compute.manager [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Detaching volume ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5 [ 989.826035] env[62000]: INFO nova.virt.block_device [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Attempting to driver detach volume ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5 from mountpoint /dev/sdb [ 989.826035] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 989.826333] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201608', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'name': 'volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8a11689f-fc00-43f8-9215-8d81daa84400', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'serial': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 989.828611] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8caf0dc2-0015-48b3-a50b-d70a3e7804ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.863304] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fc8d3e-6875-4fed-8f6f-de19ebc0d713 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.873550] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78208f09-27ba-4ef9-9f2f-c1acbdada8cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.900226] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de01c03d-a8ee-4296-8f33-890e6e82ebaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.919084] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] The volume has not been displaced from its original location: [datastore1] volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5/volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 989.924486] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfiguring VM instance instance-00000044 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 989.924848] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0c133e16-e774-40e6-b769-3fff617fc999 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.945505] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 989.945505] env[62000]: value = "task-882749" [ 989.945505] env[62000]: _type = "Task" [ 989.945505] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.954736] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608b315d-c52f-4463-859f-7ce357d8a9ff {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.963831] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882749, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.982043] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 990.171657] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.780s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.177762] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.662s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.178633] env[62000]: INFO nova.compute.claims [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 990.204527] env[62000]: INFO nova.scheduler.client.report [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance ea58da8c-e4ac-4863-942d-6294cf04e6f2 [ 990.458908] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882749, 'name': ReconfigVM_Task, 'duration_secs': 0.316742} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.460277] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Reconfigured VM instance instance-00000044 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 990.465480] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edb5de0a-ba3b-4b30-bacf-db7cc533bca7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.488870] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 990.489341] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 990.489341] env[62000]: value = "task-882751" [ 990.489341] env[62000]: _type = "Task" [ 990.489341] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.489646] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-dfed9e34-45ba-402e-b832-b0c254e22338 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.544812] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882751, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.544812] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 990.544812] env[62000]: value = "task-882752" [ 990.544812] env[62000]: _type = "Task" [ 990.544812] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.544812] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882752, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.565044] env[62000]: DEBUG nova.network.neutron [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updated VIF entry in instance network info cache for port 40a0a660-2c5f-4d14-829e-1c85ff3fc2bb. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 990.565044] env[62000]: DEBUG nova.network.neutron [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updating instance_info_cache with network_info: [{"id": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "address": "fa:16:3e:85:86:31", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40a0a660-2c", "ovs_interfaceid": "40a0a660-2c5f-4d14-829e-1c85ff3fc2bb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.715922] env[62000]: DEBUG oslo_concurrency.lockutils [None req-326bab3f-c1de-429e-8327-50e2b5041458 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "ea58da8c-e4ac-4863-942d-6294cf04e6f2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.666s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.735285] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "a2166191-09b7-4e5b-9cca-521f76814fb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.735646] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.002807] env[62000]: DEBUG oslo_vmware.api [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882751, 'name': ReconfigVM_Task, 'duration_secs': 0.155625} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.003170] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201608', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'name': 'volume-ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attached', 'instance': '8a11689f-fc00-43f8-9215-8d81daa84400', 'attached_at': '', 'detached_at': '', 'volume_id': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5', 'serial': 'ab9e39a2-95c6-41c5-a15e-61a7d6a5f8b5'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 991.017923] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882752, 'name': PowerOffVM_Task, 'duration_secs': 0.225585} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.018326] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 991.018458] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 991.067184] env[62000]: DEBUG oslo_concurrency.lockutils [req-ba51ac67-5411-4661-bd48-8f10b938746a req-a18bc85c-a679-4d33-80f4-e0bcb36a880c service nova] Releasing lock "refresh_cache-f8f26be0-f7d1-43e9-886d-c19d385e6935" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.238995] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 991.369902] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109acde9-5ec0-4b87-93f6-7828d142d416 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.379076] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63dd895-8a03-4ac0-8543-8407d2c04fc7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.412292] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f356f25-2cae-4c30-a037-e6a96ac85208 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.421433] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0f0f2fa-9809-40fe-a94c-3a01256284ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.437233] env[62000]: DEBUG nova.compute.provider_tree [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 991.527140] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.527140] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.527140] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.527140] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.527340] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.527387] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.527590] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.527746] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.527908] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.528197] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.528391] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 991.534065] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f786eaa-e540-44d1-8041-2892a42f2d9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.550793] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 991.550793] env[62000]: value = "task-882753" [ 991.550793] env[62000]: _type = "Task" [ 991.550793] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.560161] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882753, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.561993] env[62000]: DEBUG nova.objects.instance [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'flavor' on Instance uuid 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.771394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.819486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.819732] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.940804] env[62000]: DEBUG nova.scheduler.client.report [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 992.067441] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882753, 'name': ReconfigVM_Task, 'duration_secs': 0.452467} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.068946] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 992.282437] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.282688] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.322728] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.447900] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.447900] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 992.450505] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.136s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.450627] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.450823] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 992.451522] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.392s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.451522] env[62000]: DEBUG nova.objects.instance [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'resources' on Instance uuid c9758d78-83a0-430e-a69e-e2d82f99daa9 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.453474] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26c75771-9024-415d-a2fc-c4b61346f6db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.461863] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba2aec6-1767-40e7-8f14-56e76369d170 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.478660] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e25da3a2-a727-427d-b3d4-596a891b9f6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.485652] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6847f5-a321-4d70-97fb-063b7974c341 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.517229] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179816MB free_disk=52GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 992.517402] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.573545] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c16ba793-c9c7-44c9-8664-8feb4ebda51c tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.304s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.576650] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 992.576875] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 992.577279] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 992.577499] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 992.577655] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 992.577826] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 992.578061] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 992.578397] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 992.578611] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 992.578786] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 992.578964] env[62000]: DEBUG nova.virt.hardware [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.585930] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 992.587685] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f9f08c64-d129-42c9-bbf9-d2df51168469 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.611644] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 992.611644] env[62000]: value = "task-882754" [ 992.611644] env[62000]: _type = "Task" [ 992.611644] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.624667] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882754, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.787139] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 992.852575] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.954322] env[62000]: DEBUG nova.compute.utils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.956126] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.956205] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 993.032822] env[62000]: DEBUG nova.policy [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d0b65722357449ab64736e5e1eb713f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ce3270692934d2c9c1330a45c0e059e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 993.124039] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882754, 'name': ReconfigVM_Task, 'duration_secs': 0.197018} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.129209] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 993.129547] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94c512d3-643a-4989-82c4-cb937992790e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.164437] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.172667] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-98911de6-fa3c-45ec-889e-f4197f9bc81a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.190579] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 993.190579] env[62000]: value = "task-882755" [ 993.190579] env[62000]: _type = "Task" [ 993.190579] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.201983] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882755, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.210774] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1ea1df-35da-4f85-9525-bb93e09c21e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.220086] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30560fad-aa5d-4843-9095-710194bf08a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.253853] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a63f5ce-ab62-4332-bd3e-6f10fc4a624f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.262407] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696d0450-1267-4518-9835-3d374059082a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.282645] env[62000]: DEBUG nova.compute.provider_tree [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 993.311531] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.459285] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 993.524816] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Successfully created port: 40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 993.702155] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882755, 'name': ReconfigVM_Task, 'duration_secs': 0.304428} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.702512] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 993.702896] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 993.771734] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.772048] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.772310] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 993.772561] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.772816] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.780971] env[62000]: INFO nova.compute.manager [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Terminating instance [ 993.780971] env[62000]: DEBUG nova.compute.manager [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 993.780971] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 993.782230] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c18e9827-2674-4979-be89-b3718af57d56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.786572] env[62000]: DEBUG nova.scheduler.client.report [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.793678] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 993.793678] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7b01b219-f393-41b1-817c-8570f6eb7bfe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.798400] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 993.798400] env[62000]: value = "task-882757" [ 993.798400] env[62000]: _type = "Task" [ 993.798400] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.807328] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882757, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.210746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f07772-818e-405e-8a0a-259ed8313465 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.233402] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba98e14-89a8-488e-b968-d4e0daefcf82 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.253219] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 994.294370] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.843s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.297120] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.798s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.299044] env[62000]: INFO nova.compute.claims [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.309942] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882757, 'name': PowerOffVM_Task, 'duration_secs': 0.38193} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.310241] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 994.311064] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 994.311064] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f0c6e881-9d5a-4e2a-9e5a-ab290c05a370 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.322788] env[62000]: INFO nova.scheduler.client.report [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance c9758d78-83a0-430e-a69e-e2d82f99daa9 [ 994.380011] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 994.380270] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 994.380464] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] 8a11689f-fc00-43f8-9215-8d81daa84400 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 994.380723] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e35d329-cb44-4de9-8a01-17bb0b85f8cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.386671] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 994.386671] env[62000]: value = "task-882759" [ 994.386671] env[62000]: _type = "Task" [ 994.386671] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.395444] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882759, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.469348] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 994.496643] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 994.496897] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 994.497069] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 994.497262] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 994.497412] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 994.497561] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 994.497769] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 994.497929] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 994.498115] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 994.498284] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 994.498458] env[62000]: DEBUG nova.virt.hardware [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 994.499347] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1861bc2f-e1af-4efc-a322-b53212695f9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.507535] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a25a52-3d3d-45a5-8a0c-625afeeee907 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.832778] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1c0c1f01-7145-4638-86e5-27173a67b8b5 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "c9758d78-83a0-430e-a69e-e2d82f99daa9" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.348s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.897057] env[62000]: DEBUG oslo_vmware.api [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882759, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.433378} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.897348] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 994.897545] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 994.897724] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 994.897935] env[62000]: INFO nova.compute.manager [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Took 1.12 seconds to destroy the instance on the hypervisor. [ 994.898219] env[62000]: DEBUG oslo.service.loopingcall [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 994.898417] env[62000]: DEBUG nova.compute.manager [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 994.898515] env[62000]: DEBUG nova.network.neutron [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 994.982314] env[62000]: DEBUG nova.compute.manager [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Received event network-vif-plugged-40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 994.982562] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.982833] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] Lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.983114] env[62000]: DEBUG oslo_concurrency.lockutils [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] Lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.983345] env[62000]: DEBUG nova.compute.manager [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] No waiting events found dispatching network-vif-plugged-40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.984151] env[62000]: WARNING nova.compute.manager [req-7f15f0b0-ad02-4d41-bdde-25722a5100cb req-e63a112a-e4fc-44f0-a83e-89e7c264006f service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Received unexpected event network-vif-plugged-40d67815-d1be-4bab-8b61-bf89d3d62589 for instance with vm_state building and task_state spawning. [ 995.165723] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Successfully updated port: 40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 995.668315] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d76cb3-29e2-48a4-ba86-dd9e988278d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.672354] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.672672] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.672975] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 995.680128] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e9e641c-3fe3-4dfe-befb-75c0fc566e8d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.714693] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f21ec20-f3a9-4ddb-8fbc-8b6d7fad36e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.721586] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16906c58-b2e8-4aa9-bf67-c1d8601272de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.736495] env[62000]: DEBUG nova.compute.provider_tree [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.927172] env[62000]: DEBUG nova.network.neutron [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 995.997923] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.998171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.011238] env[62000]: DEBUG nova.network.neutron [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.212187] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 996.239365] env[62000]: DEBUG nova.scheduler.client.report [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.351809] env[62000]: DEBUG nova.network.neutron [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updating instance_info_cache with network_info: [{"id": "40d67815-d1be-4bab-8b61-bf89d3d62589", "address": "fa:16:3e:b5:9a:e4", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d67815-d1", "ovs_interfaceid": "40d67815-d1be-4bab-8b61-bf89d3d62589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.501604] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 996.513662] env[62000]: INFO nova.compute.manager [-] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Took 1.62 seconds to deallocate network for instance. [ 996.745787] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.745787] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 996.748467] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.977s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.750035] env[62000]: INFO nova.compute.claims [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.858020] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.858020] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Instance network_info: |[{"id": "40d67815-d1be-4bab-8b61-bf89d3d62589", "address": "fa:16:3e:b5:9a:e4", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d67815-d1", "ovs_interfaceid": "40d67815-d1be-4bab-8b61-bf89d3d62589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.858020] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b5:9a:e4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '40d67815-d1be-4bab-8b61-bf89d3d62589', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.863782] env[62000]: DEBUG oslo.service.loopingcall [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.864123] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 996.864485] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0f029669-1294-4cda-8d96-d5cf55da42f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.885111] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.885111] env[62000]: value = "task-882761" [ 996.885111] env[62000]: _type = "Task" [ 996.885111] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.893614] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882761, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.949354] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 996.950752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.950752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.019436] env[62000]: DEBUG nova.compute.manager [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Received event network-changed-40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 997.019716] env[62000]: DEBUG nova.compute.manager [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Refreshing instance network info cache due to event network-changed-40d67815-d1be-4bab-8b61-bf89d3d62589. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 997.019985] env[62000]: DEBUG oslo_concurrency.lockutils [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] Acquiring lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.020163] env[62000]: DEBUG oslo_concurrency.lockutils [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] Acquired lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.020372] env[62000]: DEBUG nova.network.neutron [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Refreshing network info cache for port 40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 997.022400] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.031910] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.254191] env[62000]: DEBUG nova.compute.utils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 997.257920] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 997.258101] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 997.296586] env[62000]: DEBUG nova.policy [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7a4e8327d3a3423185590df540e7e785', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6f27fa265194257995183660c0427b1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 997.394719] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882761, 'name': CreateVM_Task, 'duration_secs': 0.285863} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.394894] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 997.395977] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.396170] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.396494] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.396770] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-05da867d-c5d1-466e-967f-c4caaee71966 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.401186] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 997.401186] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5282ef09-b625-133b-2533-0d0b6303c492" [ 997.401186] env[62000]: _type = "Task" [ 997.401186] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.409932] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5282ef09-b625-133b-2533-0d0b6303c492, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.591717] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Successfully created port: ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 997.759104] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 997.908117] env[62000]: DEBUG nova.network.neutron [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updated VIF entry in instance network info cache for port 40d67815-d1be-4bab-8b61-bf89d3d62589. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 997.910187] env[62000]: DEBUG nova.network.neutron [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updating instance_info_cache with network_info: [{"id": "40d67815-d1be-4bab-8b61-bf89d3d62589", "address": "fa:16:3e:b5:9a:e4", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d67815-d1", "ovs_interfaceid": "40d67815-d1be-4bab-8b61-bf89d3d62589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.921785] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5282ef09-b625-133b-2533-0d0b6303c492, 'name': SearchDatastore_Task, 'duration_secs': 0.009661} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.922134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.922387] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.922639] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.923451] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.923451] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.923451] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7c320a0-ac55-414a-837a-4f3e5a084a90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.931745] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.931945] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 997.932976] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b92ff657-4581-4839-80c1-1f3d20ccba30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.945820] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 997.945820] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52611db9-3cb4-013c-c630-827905f8f655" [ 997.945820] env[62000]: _type = "Task" [ 997.945820] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.953582] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52611db9-3cb4-013c-c630-827905f8f655, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.957678] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068c45d1-ca66-45a6-ae54-c303c49bc008 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.964837] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5573bc3-00c4-4017-a003-c89e82d89cb7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.995987] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec56fbe2-a79a-4578-ab0a-38057c899b50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.002963] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be7191b3-6885-438c-88df-99ed110baf1b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.007358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 998.007527] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 998.007704] env[62000]: DEBUG nova.network.neutron [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 998.019505] env[62000]: DEBUG nova.compute.provider_tree [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 998.415628] env[62000]: DEBUG oslo_concurrency.lockutils [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] Releasing lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.415917] env[62000]: DEBUG nova.compute.manager [req-6473500d-c513-4abe-b198-735452bda3d9 req-155cd899-2015-46a2-acc3-94a21dd219b0 service nova] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Received event network-vif-deleted-7a9e2ced-47e0-4b04-8f3f-87b1e03bc0e2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 998.456206] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52611db9-3cb4-013c-c630-827905f8f655, 'name': SearchDatastore_Task, 'duration_secs': 0.007781} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.456959] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1c5abbbb-cff4-4dbf-9332-26a125fd623c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.461735] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 998.461735] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5279405a-6121-1b82-8238-893d1b830816" [ 998.461735] env[62000]: _type = "Task" [ 998.461735] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.468787] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5279405a-6121-1b82-8238-893d1b830816, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.521546] env[62000]: DEBUG nova.scheduler.client.report [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.721504] env[62000]: DEBUG nova.network.neutron [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.774691] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 998.800071] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 998.800361] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 998.800523] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 998.800707] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 998.800857] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 998.801014] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 998.801290] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 998.801404] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 998.801573] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 998.801739] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 998.801912] env[62000]: DEBUG nova.virt.hardware [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 998.803183] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc682d7-48a8-463a-844a-a64433043654 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.811195] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-142d1b17-21a2-4f57-8ae5-134da1db39af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.974053] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5279405a-6121-1b82-8238-893d1b830816, 'name': SearchDatastore_Task, 'duration_secs': 0.008791} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.974379] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.974546] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0246b032-505e-4bc5-bfc0-5779ff564626/0246b032-505e-4bc5-bfc0-5779ff564626.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 998.974809] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-efa44a36-1f24-446d-aa7c-692d8075a5de {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.981527] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 998.981527] env[62000]: value = "task-882763" [ 998.981527] env[62000]: _type = "Task" [ 998.981527] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.989897] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882763, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.027655] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.279s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.027655] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 999.030554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.513s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.224486] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 999.271851] env[62000]: DEBUG nova.compute.manager [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Received event network-vif-plugged-ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 999.272176] env[62000]: DEBUG oslo_concurrency.lockutils [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] Acquiring lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 999.272441] env[62000]: DEBUG oslo_concurrency.lockutils [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 999.272538] env[62000]: DEBUG oslo_concurrency.lockutils [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 999.272754] env[62000]: DEBUG nova.compute.manager [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] No waiting events found dispatching network-vif-plugged-ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 999.272885] env[62000]: WARNING nova.compute.manager [req-878fc56b-9231-4cd5-ba3a-d00dd196d470 req-56f09fba-abf1-4803-8c67-460050acf9f2 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Received unexpected event network-vif-plugged-ad177e0f-5b23-4aa7-87ae-52f9de2e6650 for instance with vm_state building and task_state spawning. [ 999.374534] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Successfully updated port: ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 999.491494] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882763, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476516} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.491769] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 0246b032-505e-4bc5-bfc0-5779ff564626/0246b032-505e-4bc5-bfc0-5779ff564626.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 999.491996] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.492296] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbc4c5f1-dbe4-43d5-bf34-be8fb55d21a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.498790] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 999.498790] env[62000]: value = "task-882764" [ 999.498790] env[62000]: _type = "Task" [ 999.498790] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.506064] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.532932] env[62000]: DEBUG nova.compute.utils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 999.540154] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 999.540358] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 999.581316] env[62000]: DEBUG nova.policy [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '68969069b2bf4e2c83d25083ce5e577e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0b241049582495a9ffb67057a999451', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 999.738814] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0a2d2e-3a94-4c69-b451-191a370e00da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.748666] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00b2ba7b-732d-4afd-bafe-a467fd637f62 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.867938] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Successfully created port: c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.876744] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 999.876908] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquired lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 999.877079] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1000.010488] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.041310] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Applying migration context for instance 3878579f-6435-4fe3-9f8c-8461d8ac57ee as it has an incoming, in-progress migration 5793fe93-6df4-44ae-af50-d526a12184f6. Migration status is post-migrating {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1000.042421] env[62000]: INFO nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating resource usage from migration 5793fe93-6df4-44ae-af50-d526a12184f6 [ 1000.044896] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1000.076037] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.076196] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 44a54bd7-a35c-49ed-85ed-346830cee6ad actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.076342] env[62000]: WARNING nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 8a11689f-fc00-43f8-9215-8d81daa84400 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 1000.076461] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance f8f26be0-f7d1-43e9-886d-c19d385e6935 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.076576] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Migration 5793fe93-6df4-44ae-af50-d526a12184f6 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1000.076685] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 3878579f-6435-4fe3-9f8c-8461d8ac57ee actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.076792] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 0246b032-505e-4bc5-bfc0-5779ff564626 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.076896] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 9f864cad-f454-4dc3-b6d0-793d7adec0ae actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.077064] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance a2166191-09b7-4e5b-9cca-521f76814fb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1000.415518] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1000.509841] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.970688} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.510148] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1000.510950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21966bb8-107d-4750-800e-87379712eb24 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.535347] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 0246b032-505e-4bc5-bfc0-5779ff564626/0246b032-505e-4bc5-bfc0-5779ff564626.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1000.535647] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b37ce1e9-5d86-48ea-8296-658b0bc4787c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.568708] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1000.568708] env[62000]: value = "task-882766" [ 1000.568708] env[62000]: _type = "Task" [ 1000.568708] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.574421] env[62000]: DEBUG nova.network.neutron [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Updating instance_info_cache with network_info: [{"id": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "address": "fa:16:3e:ae:f9:17", "network": {"id": "2fad45ed-5c97-4aee-a275-863d19d02c64", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-46653558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d6f27fa265194257995183660c0427b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad177e0f-5b", "ovs_interfaceid": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1000.578802] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882766, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.579763] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 9ccf07fa-90db-40b9-af5d-b47ee5773f9c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1000.858373] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20048fa-7ba3-4f8b-9e9b-2bf46f3a0fd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.882400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be43f74-398d-4566-9511-bfa2324187cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.889754] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1001.064239] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1001.079199] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882766, 'name': ReconfigVM_Task, 'duration_secs': 0.282321} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.079716] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Releasing lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.080256] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Instance network_info: |[{"id": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "address": "fa:16:3e:ae:f9:17", "network": {"id": "2fad45ed-5c97-4aee-a275-863d19d02c64", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-46653558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d6f27fa265194257995183660c0427b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad177e0f-5b", "ovs_interfaceid": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1001.080256] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 0246b032-505e-4bc5-bfc0-5779ff564626/0246b032-505e-4bc5-bfc0-5779ff564626.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1001.081051] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ae:f9:17', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fe20ef0e-0991-44d7-887d-08dddac0b56b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad177e0f-5b23-4aa7-87ae-52f9de2e6650', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1001.088721] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Creating folder: Project (d6f27fa265194257995183660c0427b1). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1001.088971] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5b3d921a-a205-48c9-b4ec-97cb42f6eff2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.093046] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 64ca26bd-dc8c-4f00-bfde-a24f8d650848 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1001.094286] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-62243a14-823e-4ed8-8a21-1c3234290f34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.102786] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1001.102786] env[62000]: value = "task-882767" [ 1001.102786] env[62000]: _type = "Task" [ 1001.102786] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.104498] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.104732] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.104918] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.105164] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.105336] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.105490] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.105699] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.105893] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.106093] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.106269] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.106494] env[62000]: DEBUG nova.virt.hardware [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.107334] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ccc73c-0903-41ee-ae50-fbd02ff003b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.115151] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Created folder: Project (d6f27fa265194257995183660c0427b1) in parent group-v201431. [ 1001.115330] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Creating folder: Instances. Parent ref: group-v201634. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1001.116248] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-526b6685-3be3-4460-acb5-81feff4c3fd6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.126642] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9798464-1f83-4dcf-8259-aa0b64770136 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.131011] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882767, 'name': Rename_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.132487] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Created folder: Instances in parent group-v201634. [ 1001.132781] env[62000]: DEBUG oslo.service.loopingcall [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1001.133377] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1001.133674] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9bfe4f77-3237-4de6-8f3b-6e0b745f8d73 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.164811] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1001.164811] env[62000]: value = "task-882770" [ 1001.164811] env[62000]: _type = "Task" [ 1001.164811] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.173151] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882770, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.303613] env[62000]: DEBUG nova.compute.manager [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Received event network-changed-ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1001.303613] env[62000]: DEBUG nova.compute.manager [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Refreshing instance network info cache due to event network-changed-ad177e0f-5b23-4aa7-87ae-52f9de2e6650. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1001.303613] env[62000]: DEBUG oslo_concurrency.lockutils [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] Acquiring lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.303613] env[62000]: DEBUG oslo_concurrency.lockutils [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] Acquired lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.303613] env[62000]: DEBUG nova.network.neutron [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Refreshing network info cache for port ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1001.397367] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1001.397851] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf59707d-d896-47c0-9ece-a9d42f14002a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.406021] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1001.406021] env[62000]: value = "task-882771" [ 1001.406021] env[62000]: _type = "Task" [ 1001.406021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.414991] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882771, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.597985] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 67f8274b-a0ad-419a-81fc-515b06ad41aa has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1001.598316] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1001.598544] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1001.613167] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Successfully updated port: c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1001.617430] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882767, 'name': Rename_Task, 'duration_secs': 0.171358} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.620634] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1001.621156] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd520442-c1c0-47f7-af99-36f78a3a8988 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.630429] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1001.630429] env[62000]: value = "task-882773" [ 1001.630429] env[62000]: _type = "Task" [ 1001.630429] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.643358] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882773, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.676605] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882770, 'name': CreateVM_Task, 'duration_secs': 0.385254} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.676605] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1001.676895] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.677021] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.677416] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1001.677680] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1543f4ea-9ed6-4694-97ec-83aba2127e4f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.682440] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1001.682440] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52965892-52c7-081c-becf-1b18b4441c51" [ 1001.682440] env[62000]: _type = "Task" [ 1001.682440] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.693625] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52965892-52c7-081c-becf-1b18b4441c51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.795641] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfcdd032-6700-45b8-9849-500adc9ee099 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.804509] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f04417a6-6635-4b0a-93ad-ec78738c6f57 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.847810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f054e4d-543f-4b7c-af57-e59f45c488b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.858490] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62d21fb5-a712-4a7e-87e2-472be2b174f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.873553] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.916546] env[62000]: DEBUG oslo_vmware.api [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882771, 'name': PowerOnVM_Task, 'duration_secs': 0.486943} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.917166] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1001.917372] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3fc39ebb-48d9-4c6f-bef4-2184d9adfafc tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance '3878579f-6435-4fe3-9f8c-8461d8ac57ee' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1002.054318] env[62000]: DEBUG nova.network.neutron [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Updated VIF entry in instance network info cache for port ad177e0f-5b23-4aa7-87ae-52f9de2e6650. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1002.054853] env[62000]: DEBUG nova.network.neutron [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Updating instance_info_cache with network_info: [{"id": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "address": "fa:16:3e:ae:f9:17", "network": {"id": "2fad45ed-5c97-4aee-a275-863d19d02c64", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-46653558-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "d6f27fa265194257995183660c0427b1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fe20ef0e-0991-44d7-887d-08dddac0b56b", "external-id": "nsx-vlan-transportzone-991", "segmentation_id": 991, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad177e0f-5b", "ovs_interfaceid": "ad177e0f-5b23-4aa7-87ae-52f9de2e6650", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.121286] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.121686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquired lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.121686] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1002.141706] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882773, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.193272] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52965892-52c7-081c-becf-1b18b4441c51, 'name': SearchDatastore_Task, 'duration_secs': 0.028153} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.193457] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.193702] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1002.193950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.194125] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.194348] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1002.194617] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ae6160cc-a2bc-42e6-acc8-64c325225b34 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.205489] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1002.205682] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1002.206509] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4f9e8a66-98d3-4a61-b849-2dff5817d45f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.212313] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1002.212313] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cc1f90-4296-dcc8-ec77-c8c7fe36f0d0" [ 1002.212313] env[62000]: _type = "Task" [ 1002.212313] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.220595] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cc1f90-4296-dcc8-ec77-c8c7fe36f0d0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.376902] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.558341] env[62000]: DEBUG oslo_concurrency.lockutils [req-8eb2771f-feac-48d4-b3fc-d17069a1ab1a req-388f241f-20a2-44f1-9f41-21f80e0cef51 service nova] Releasing lock "refresh_cache-9f864cad-f454-4dc3-b6d0-793d7adec0ae" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.642945] env[62000]: DEBUG oslo_vmware.api [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882773, 'name': PowerOnVM_Task, 'duration_secs': 0.954656} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.643286] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1002.643906] env[62000]: INFO nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Took 8.17 seconds to spawn the instance on the hypervisor. [ 1002.643906] env[62000]: DEBUG nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.644727] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b800282-4d41-4710-8d76-f91286708296 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.653659] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1002.722809] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52cc1f90-4296-dcc8-ec77-c8c7fe36f0d0, 'name': SearchDatastore_Task, 'duration_secs': 0.016919} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.726148] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3608e7ac-3b35-4961-84e7-ecedf543a5e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.731756] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1002.731756] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c548f4-bce3-c170-46a1-9d44697583f6" [ 1002.731756] env[62000]: _type = "Task" [ 1002.731756] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.738922] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c548f4-bce3-c170-46a1-9d44697583f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.799678] env[62000]: DEBUG nova.network.neutron [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Updating instance_info_cache with network_info: [{"id": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "address": "fa:16:3e:aa:7e:ef", "network": {"id": "97f18c18-005d-47d0-a223-b0f836afbbc7", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1264660438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0b241049582495a9ffb67057a999451", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1ca3821-11", "ovs_interfaceid": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.882144] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1002.882293] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.852s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.882924] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.031s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1002.884524] env[62000]: INFO nova.compute.claims [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1003.160928] env[62000]: INFO nova.compute.manager [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Took 20.66 seconds to build instance. [ 1003.242545] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c548f4-bce3-c170-46a1-9d44697583f6, 'name': SearchDatastore_Task, 'duration_secs': 0.015267} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.242816] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.243394] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9f864cad-f454-4dc3-b6d0-793d7adec0ae/9f864cad-f454-4dc3-b6d0-793d7adec0ae.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1003.243394] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b174d0c1-3544-40b0-adb0-509ed0eec165 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.249957] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1003.249957] env[62000]: value = "task-882774" [ 1003.249957] env[62000]: _type = "Task" [ 1003.249957] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.257415] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882774, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.302201] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Releasing lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.302535] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Instance network_info: |[{"id": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "address": "fa:16:3e:aa:7e:ef", "network": {"id": "97f18c18-005d-47d0-a223-b0f836afbbc7", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1264660438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0b241049582495a9ffb67057a999451", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1ca3821-11", "ovs_interfaceid": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1003.303009] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:7e:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'bec1528b-3e87-477b-8ab2-02696ad47e66', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c1ca3821-1188-40f4-9056-a9d3b36b7a00', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1003.310633] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Creating folder: Project (f0b241049582495a9ffb67057a999451). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1003.310795] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d4f0474b-c0dd-4a63-8ed9-a8a9722da76a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.322927] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Created folder: Project (f0b241049582495a9ffb67057a999451) in parent group-v201431. [ 1003.323158] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Creating folder: Instances. Parent ref: group-v201637. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1003.323469] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bfcc6044-a96b-4c9c-abf1-4faf30c5d6bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.332508] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Created folder: Instances in parent group-v201637. [ 1003.332767] env[62000]: DEBUG oslo.service.loopingcall [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1003.332973] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1003.333255] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-368dc4bf-8871-49d2-b2a9-af932bcc264c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.352854] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1003.352854] env[62000]: value = "task-882777" [ 1003.352854] env[62000]: _type = "Task" [ 1003.352854] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.360409] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882777, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.377017] env[62000]: DEBUG nova.compute.manager [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Received event network-vif-plugged-c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.377017] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Acquiring lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.377017] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.377017] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.377426] env[62000]: DEBUG nova.compute.manager [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] No waiting events found dispatching network-vif-plugged-c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.377426] env[62000]: WARNING nova.compute.manager [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Received unexpected event network-vif-plugged-c1ca3821-1188-40f4-9056-a9d3b36b7a00 for instance with vm_state building and task_state spawning. [ 1003.377515] env[62000]: DEBUG nova.compute.manager [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Received event network-changed-c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1003.377652] env[62000]: DEBUG nova.compute.manager [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Refreshing instance network info cache due to event network-changed-c1ca3821-1188-40f4-9056-a9d3b36b7a00. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1003.377829] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Acquiring lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.377965] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Acquired lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.378137] env[62000]: DEBUG nova.network.neutron [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Refreshing network info cache for port c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1003.663110] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c513b59e-f3cb-481a-b9c9-098f5593d9fc tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.174s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.760516] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882774, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.496418} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.760799] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9f864cad-f454-4dc3-b6d0-793d7adec0ae/9f864cad-f454-4dc3-b6d0-793d7adec0ae.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1003.761101] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1003.761289] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6e40b21-4c24-4e3b-b19b-ee467c5bbe03 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.768380] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1003.768380] env[62000]: value = "task-882779" [ 1003.768380] env[62000]: _type = "Task" [ 1003.768380] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.777649] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882779, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.863523] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882777, 'name': CreateVM_Task, 'duration_secs': 0.379442} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.863523] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1003.864210] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.864389] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.864722] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.864984] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-18d70258-3100-4e7f-8195-78d1dd34a089 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.869630] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1003.869630] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52257402-2871-5c71-b2c1-ba2485df7a0c" [ 1003.869630] env[62000]: _type = "Task" [ 1003.869630] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.877409] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52257402-2871-5c71-b2c1-ba2485df7a0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.049447] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6205748-caae-453b-9479-d39de27c3b7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.059455] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b673d106-cfae-40df-b206-ec0e9b34c04e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.089372] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fbae188-7fcd-4f09-b645-80fe71612d98 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.096944] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c68b6d2-5954-4b89-8492-731a58403c33 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.111564] env[62000]: DEBUG nova.compute.provider_tree [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1004.272876] env[62000]: DEBUG nova.network.neutron [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Updated VIF entry in instance network info cache for port c1ca3821-1188-40f4-9056-a9d3b36b7a00. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1004.273273] env[62000]: DEBUG nova.network.neutron [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Updating instance_info_cache with network_info: [{"id": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "address": "fa:16:3e:aa:7e:ef", "network": {"id": "97f18c18-005d-47d0-a223-b0f836afbbc7", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-1264660438-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0b241049582495a9ffb67057a999451", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "bec1528b-3e87-477b-8ab2-02696ad47e66", "external-id": "nsx-vlan-transportzone-180", "segmentation_id": 180, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc1ca3821-11", "ovs_interfaceid": "c1ca3821-1188-40f4-9056-a9d3b36b7a00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.280152] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882779, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064181} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.280453] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1004.281610] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33156229-47a2-47cd-bec9-2907d23a0c81 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.310560] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 9f864cad-f454-4dc3-b6d0-793d7adec0ae/9f864cad-f454-4dc3-b6d0-793d7adec0ae.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1004.311184] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4524e02a-ebe2-4b99-b4ac-0e651f62cf76 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.327592] env[62000]: DEBUG nova.network.neutron [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Port 1eb9c5c5-8fe1-4652-986f-3b9abdd36485 binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1004.327872] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.328045] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.328224] env[62000]: DEBUG nova.network.neutron [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1004.337122] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1004.337122] env[62000]: value = "task-882780" [ 1004.337122] env[62000]: _type = "Task" [ 1004.337122] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.346503] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882780, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.381586] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52257402-2871-5c71-b2c1-ba2485df7a0c, 'name': SearchDatastore_Task, 'duration_secs': 0.009775} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.381880] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.382245] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1004.382415] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.382614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.382822] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1004.383073] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1ec3e3da-66f4-4e82-91e0-81aae6d33a47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.393321] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1004.393517] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1004.394306] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0a7f10ac-708e-4d31-b514-5fa928690af1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.400066] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1004.400066] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52537b03-75d6-46ae-dec1-eda3b645fe9e" [ 1004.400066] env[62000]: _type = "Task" [ 1004.400066] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.408680] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52537b03-75d6-46ae-dec1-eda3b645fe9e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.616454] env[62000]: DEBUG nova.scheduler.client.report [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1004.776245] env[62000]: DEBUG oslo_concurrency.lockutils [req-3df4932f-9431-4011-80e0-8ced056ef603 req-316934ad-1261-4b10-8487-fa5f931ef58e service nova] Releasing lock "refresh_cache-a2166191-09b7-4e5b-9cca-521f76814fb2" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.851678] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882780, 'name': ReconfigVM_Task, 'duration_secs': 0.362629} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.851678] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 9f864cad-f454-4dc3-b6d0-793d7adec0ae/9f864cad-f454-4dc3-b6d0-793d7adec0ae.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1004.851678] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6c382988-c785-4847-800f-0920cc1e1911 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.858476] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1004.858476] env[62000]: value = "task-882781" [ 1004.858476] env[62000]: _type = "Task" [ 1004.858476] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.869316] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882781, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.910610] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52537b03-75d6-46ae-dec1-eda3b645fe9e, 'name': SearchDatastore_Task, 'duration_secs': 0.022719} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.911401] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b537b3e6-210b-4103-b87b-091e9f00fd7d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.917594] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1004.917594] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205bce5-5c87-93a6-b1ee-3e0daba50f7c" [ 1004.917594] env[62000]: _type = "Task" [ 1004.917594] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.926638] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205bce5-5c87-93a6-b1ee-3e0daba50f7c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.098237] env[62000]: DEBUG nova.network.neutron [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.121848] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.122453] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1005.125166] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.814s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.126595] env[62000]: INFO nova.compute.claims [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1005.370806] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882781, 'name': Rename_Task, 'duration_secs': 0.362374} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.370806] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1005.370806] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-16c10cfd-5d12-4620-a00a-6eb101e5feac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.378639] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1005.378639] env[62000]: value = "task-882782" [ 1005.378639] env[62000]: _type = "Task" [ 1005.378639] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.386707] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882782, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.402898] env[62000]: DEBUG nova.compute.manager [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Received event network-changed-40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1005.403101] env[62000]: DEBUG nova.compute.manager [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Refreshing instance network info cache due to event network-changed-40d67815-d1be-4bab-8b61-bf89d3d62589. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1005.403344] env[62000]: DEBUG oslo_concurrency.lockutils [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] Acquiring lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1005.403492] env[62000]: DEBUG oslo_concurrency.lockutils [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] Acquired lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1005.403660] env[62000]: DEBUG nova.network.neutron [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Refreshing network info cache for port 40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1005.428632] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205bce5-5c87-93a6-b1ee-3e0daba50f7c, 'name': SearchDatastore_Task, 'duration_secs': 0.06195} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.428887] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.429265] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a2166191-09b7-4e5b-9cca-521f76814fb2/a2166191-09b7-4e5b-9cca-521f76814fb2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1005.429561] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-17f46665-c3af-4234-b205-21dede3ede14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.436583] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1005.436583] env[62000]: value = "task-882783" [ 1005.436583] env[62000]: _type = "Task" [ 1005.436583] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.445300] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882783, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.603224] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1005.631505] env[62000]: DEBUG nova.compute.utils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1005.634690] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1005.634910] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1005.679022] env[62000]: DEBUG nova.policy [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '674db9508719490ba7e22a3c5108837a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6742b3d48eb945fe8a11d0ee651abd1a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1005.891430] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882782, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.948114] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882783, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.106858] env[62000]: DEBUG nova.compute.manager [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62000) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1006.141765] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1006.197290] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Successfully created port: 5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1006.309518] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e77140f-3d96-4d9e-af1c-c2dd37a0aced {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.317807] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0086ce-320b-4581-8b33-57f004cecde0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.351274] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e013da-8d62-452b-a19d-a3499fc6475b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.359243] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae698182-8054-475c-a799-4f44591b4900 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.373648] env[62000]: DEBUG nova.compute.provider_tree [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1006.389830] env[62000]: DEBUG oslo_vmware.api [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882782, 'name': PowerOnVM_Task, 'duration_secs': 0.582823} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.390145] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1006.390358] env[62000]: INFO nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Took 7.62 seconds to spawn the instance on the hypervisor. [ 1006.390695] env[62000]: DEBUG nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1006.391467] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9159f002-9f2b-436f-bd3d-c658aba72582 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.413534] env[62000]: DEBUG nova.network.neutron [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updated VIF entry in instance network info cache for port 40d67815-d1be-4bab-8b61-bf89d3d62589. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1006.413871] env[62000]: DEBUG nova.network.neutron [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updating instance_info_cache with network_info: [{"id": "40d67815-d1be-4bab-8b61-bf89d3d62589", "address": "fa:16:3e:b5:9a:e4", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap40d67815-d1", "ovs_interfaceid": "40d67815-d1be-4bab-8b61-bf89d3d62589", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.447700] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882783, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71337} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.448044] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] a2166191-09b7-4e5b-9cca-521f76814fb2/a2166191-09b7-4e5b-9cca-521f76814fb2.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1006.448281] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1006.448540] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5bb605f-2cdb-4279-a88c-1f918fd4ed55 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.455369] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1006.455369] env[62000]: value = "task-882785" [ 1006.455369] env[62000]: _type = "Task" [ 1006.455369] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.463632] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882785, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.879285] env[62000]: DEBUG nova.scheduler.client.report [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1006.908934] env[62000]: INFO nova.compute.manager [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Took 17.43 seconds to build instance. [ 1006.916585] env[62000]: DEBUG oslo_concurrency.lockutils [req-fcd4efc2-6de8-49b8-91be-c17ccf553387 req-0cde7a89-9580-4ff5-96b8-43fe3058d896 service nova] Releasing lock "refresh_cache-0246b032-505e-4bc5-bfc0-5779ff564626" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.966980] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882785, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.310023} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.967293] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.968095] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-183315b5-e087-48c1-bb96-2d1f3a571fd3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.991143] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] a2166191-09b7-4e5b-9cca-521f76814fb2/a2166191-09b7-4e5b-9cca-521f76814fb2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.991737] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ad1b28d9-a87d-492f-9860-32398166f00f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.011371] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1007.011371] env[62000]: value = "task-882786" [ 1007.011371] env[62000]: _type = "Task" [ 1007.011371] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.019674] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.157036] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1007.188912] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1007.189202] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1007.189375] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1007.189705] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1007.189788] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1007.189952] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1007.190633] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1007.190633] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1007.190633] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1007.190753] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1007.190934] env[62000]: DEBUG nova.virt.hardware [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1007.191816] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b698d073-ae46-4afa-ac8c-dbaca9612033 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.200630] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade65676-4901-4d18-98dc-1606e20d18df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.239330] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.384578] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.385337] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1007.388091] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 10.366s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.388303] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.390381] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.359s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.392506] env[62000]: INFO nova.compute.claims [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1007.410920] env[62000]: DEBUG oslo_concurrency.lockutils [None req-54115a55-8ba4-4021-9eb6-83e6a5636180 tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.944s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.416944] env[62000]: INFO nova.scheduler.client.report [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted allocations for instance 8a11689f-fc00-43f8-9215-8d81daa84400 [ 1007.522352] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882786, 'name': ReconfigVM_Task, 'duration_secs': 0.410787} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.522650] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Reconfigured VM instance instance-00000061 to attach disk [datastore2] a2166191-09b7-4e5b-9cca-521f76814fb2/a2166191-09b7-4e5b-9cca-521f76814fb2.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1007.523319] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ecb1539e-d1df-418c-929d-65b8dd876d24 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.530823] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1007.530823] env[62000]: value = "task-882787" [ 1007.530823] env[62000]: _type = "Task" [ 1007.530823] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.540588] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882787, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.785705] env[62000]: DEBUG nova.compute.manager [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Received event network-vif-plugged-5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1007.786381] env[62000]: DEBUG oslo_concurrency.lockutils [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.787835] env[62000]: DEBUG oslo_concurrency.lockutils [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1007.787835] env[62000]: DEBUG oslo_concurrency.lockutils [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1007.787835] env[62000]: DEBUG nova.compute.manager [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] No waiting events found dispatching network-vif-plugged-5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1007.787835] env[62000]: WARNING nova.compute.manager [req-c6c261cc-5970-4a4a-b6f3-e7e9be86dd69 req-c9f60c06-7395-44a5-908f-8a7db9c6b9e5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Received unexpected event network-vif-plugged-5f0f0c25-4143-4468-b1e5-48897bd334ac for instance with vm_state building and task_state spawning. [ 1007.900824] env[62000]: DEBUG nova.compute.utils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1007.905864] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1007.907227] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1007.928723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-04f76653-4b63-4221-95c6-cd445e2f0125 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "8a11689f-fc00-43f8-9215-8d81daa84400" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.156s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.041947] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882787, 'name': Rename_Task, 'duration_secs': 0.143354} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.042679] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1008.043159] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94cf4c02-eebe-47a2-802a-3020d03e81a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.047479] env[62000]: DEBUG nova.policy [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1008.053021] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1008.053021] env[62000]: value = "task-882788" [ 1008.053021] env[62000]: _type = "Task" [ 1008.053021] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.060042] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882788, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.207741] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Successfully updated port: 5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.342322] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.342614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.342826] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.343134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.343250] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.346167] env[62000]: INFO nova.compute.manager [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Terminating instance [ 1008.347988] env[62000]: DEBUG nova.compute.manager [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1008.348206] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1008.349319] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a016ee9-adf2-46e7-bc5c-b8c96691d8d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.357838] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1008.358141] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-939ec7af-e8af-4322-a9fa-bbc93e4312bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.368119] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1008.368119] env[62000]: value = "task-882789" [ 1008.368119] env[62000]: _type = "Task" [ 1008.368119] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.377736] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882789, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.410746] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1008.518240] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Successfully created port: 37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1008.564448] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882788, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.609555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3fd392-546f-488a-93db-b23800fd27eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.618149] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77125542-1146-4f6d-9b08-a64d14a69087 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.650468] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4747b480-608b-4a5c-952e-14efdd3d9264 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.658646] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67db68b7-a097-4ebc-8e0d-30f70b552b8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.674069] env[62000]: DEBUG nova.compute.provider_tree [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.714024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.714024] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.714024] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1008.878191] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882789, 'name': PowerOffVM_Task, 'duration_secs': 0.219988} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.878480] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1008.878655] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1008.878920] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2681539-949e-47f8-ae11-bfc882802b8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.951863] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1008.955143] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1008.955143] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Deleting the datastore file [datastore2] 9f864cad-f454-4dc3-b6d0-793d7adec0ae {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1008.955143] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-548308e3-9117-4d05-84bd-d3eceb550c7f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.960441] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for the task: (returnval){ [ 1008.960441] env[62000]: value = "task-882791" [ 1008.960441] env[62000]: _type = "Task" [ 1008.960441] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.970550] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882791, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.064291] env[62000]: DEBUG oslo_vmware.api [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882788, 'name': PowerOnVM_Task, 'duration_secs': 0.560804} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.064834] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1009.065218] env[62000]: INFO nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Took 8.00 seconds to spawn the instance on the hypervisor. [ 1009.065550] env[62000]: DEBUG nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.066555] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-255efa6f-f669-41c6-89e2-efa8b9b9982a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.178133] env[62000]: DEBUG nova.scheduler.client.report [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1009.246875] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1009.395948] env[62000]: DEBUG nova.network.neutron [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating instance_info_cache with network_info: [{"id": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "address": "fa:16:3e:40:72:07", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f0f0c25-41", "ovs_interfaceid": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.424841] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1009.453324] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1009.453621] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1009.453792] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1009.453980] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1009.454151] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1009.454305] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1009.454551] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1009.454720] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1009.454890] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1009.455065] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1009.455244] env[62000]: DEBUG nova.virt.hardware [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1009.456180] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b0aaf5-88e4-4ecd-9d1f-b7dfb7da8e0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.467723] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3360d1-ae90-4cc0-856b-127c93c49fb3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.489114] env[62000]: DEBUG oslo_vmware.api [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Task: {'id': task-882791, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.266326} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.499622] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1009.499883] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1009.500117] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1009.500395] env[62000]: INFO nova.compute.manager [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1009.500729] env[62000]: DEBUG oslo.service.loopingcall [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.501228] env[62000]: DEBUG nova.compute.manager [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1009.501334] env[62000]: DEBUG nova.network.neutron [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1009.587568] env[62000]: INFO nova.compute.manager [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Took 17.84 seconds to build instance. [ 1009.683962] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.293s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.684526] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.687269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 2.448s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.890094] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.890325] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.898100] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1009.900662] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Instance network_info: |[{"id": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "address": "fa:16:3e:40:72:07", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f0f0c25-41", "ovs_interfaceid": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1009.900662] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:40:72:07', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f5f49a7d-c6e5-404f-b71a-91d8c070cd18', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5f0f0c25-4143-4468-b1e5-48897bd334ac', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1009.906445] env[62000]: DEBUG oslo.service.loopingcall [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1009.907499] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1009.907749] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3caf7599-f5a4-433e-b19f-57363b75feb6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.930650] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1009.930650] env[62000]: value = "task-882792" [ 1009.930650] env[62000]: _type = "Task" [ 1009.930650] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.939203] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882792, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.032822] env[62000]: DEBUG nova.compute.manager [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Received event network-changed-5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1010.033521] env[62000]: DEBUG nova.compute.manager [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Refreshing instance network info cache due to event network-changed-5f0f0c25-4143-4468-b1e5-48897bd334ac. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1010.033796] env[62000]: DEBUG oslo_concurrency.lockutils [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] Acquiring lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.033930] env[62000]: DEBUG oslo_concurrency.lockutils [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] Acquired lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.034121] env[62000]: DEBUG nova.network.neutron [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Refreshing network info cache for port 5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1010.089042] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d3c97fb9-ab54-436f-adb8-509ccde317dc tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.353s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.190915] env[62000]: DEBUG nova.compute.utils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1010.192832] env[62000]: DEBUG nova.objects.instance [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'migration_context' on Instance uuid 3878579f-6435-4fe3-9f8c-8461d8ac57ee {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.193884] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1010.194112] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1010.255863] env[62000]: DEBUG nova.policy [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1010.323037] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Successfully updated port: 37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1010.381113] env[62000]: DEBUG nova.network.neutron [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1010.394975] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1010.441718] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882792, 'name': CreateVM_Task, 'duration_secs': 0.501063} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.442096] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1010.442708] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.442894] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.443358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1010.443643] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ba29743-0c95-4326-a5c4-98ea5ffca3a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.448942] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1010.448942] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ae7f0a-1d58-6401-e5e6-f367e84d0848" [ 1010.448942] env[62000]: _type = "Task" [ 1010.448942] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.461940] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ae7f0a-1d58-6401-e5e6-f367e84d0848, 'name': SearchDatastore_Task, 'duration_secs': 0.009681} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.462270] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.462475] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1010.462713] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.462864] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.463131] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1010.463550] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-829c9c24-4f8b-496f-bad4-2188baff0c0c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.476043] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1010.476043] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1010.476043] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-577bf861-2bd3-4620-9914-9b545fd76f9a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.481082] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1010.481082] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d9331b-b328-d654-d640-9e870313ce44" [ 1010.481082] env[62000]: _type = "Task" [ 1010.481082] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.490468] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d9331b-b328-d654-d640-9e870313ce44, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.499048] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "a2166191-09b7-4e5b-9cca-521f76814fb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.499358] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.499622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.499817] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.499993] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.503454] env[62000]: INFO nova.compute.manager [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Terminating instance [ 1010.506102] env[62000]: DEBUG nova.compute.manager [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.506305] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1010.507131] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a082d1c-0a46-4ab2-953c-6f18ad445a64 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.515244] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1010.515537] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bdee0f4c-bdb2-468d-b246-9072d680fa6f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.523382] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1010.523382] env[62000]: value = "task-882793" [ 1010.523382] env[62000]: _type = "Task" [ 1010.523382] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.532889] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882793, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.543954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.544257] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.642816] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Successfully created port: e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.696186] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.825218] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.825368] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.825517] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1010.885833] env[62000]: INFO nova.compute.manager [-] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Took 1.38 seconds to deallocate network for instance. [ 1010.926775] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74cedeeb-9b3a-4bbb-9da2-4df78a446ddc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.933255] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.940704] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f69e4b6e-09d9-400c-9f7b-a4b4fdf30ea3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.976337] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c63b4c66-26bc-47a9-a577-2561dca6b80e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.983870] env[62000]: DEBUG nova.network.neutron [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updated VIF entry in instance network info cache for port 5f0f0c25-4143-4468-b1e5-48897bd334ac. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1010.984245] env[62000]: DEBUG nova.network.neutron [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating instance_info_cache with network_info: [{"id": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "address": "fa:16:3e:40:72:07", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f0f0c25-41", "ovs_interfaceid": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.000885] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebdc3ccb-aebd-438d-a938-9af410a8acaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.004986] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d9331b-b328-d654-d640-9e870313ce44, 'name': SearchDatastore_Task, 'duration_secs': 0.009624} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.006686] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-405adb7c-4ac7-4631-af8f-d45ca4899624 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.016848] env[62000]: DEBUG nova.compute.provider_tree [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.021446] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1011.021446] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52511fa8-7b8f-bc42-f742-e7465188437a" [ 1011.021446] env[62000]: _type = "Task" [ 1011.021446] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.036731] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52511fa8-7b8f-bc42-f742-e7465188437a, 'name': SearchDatastore_Task, 'duration_secs': 0.010052} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.036977] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882793, 'name': PowerOffVM_Task, 'duration_secs': 0.444983} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.037508] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.037508] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9ccf07fa-90db-40b9-af5d-b47ee5773f9c/9ccf07fa-90db-40b9-af5d-b47ee5773f9c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1011.037755] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1011.037921] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1011.038164] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-866b513d-9f70-4c89-8d5d-bc2432f5102f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.040395] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7b9cdd2a-39bb-4f79-af17-b516b48d5445 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.046835] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1011.051171] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1011.051171] env[62000]: value = "task-882794" [ 1011.051171] env[62000]: _type = "Task" [ 1011.051171] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.061826] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882794, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.113205] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1011.113205] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1011.113391] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Deleting the datastore file [datastore2] a2166191-09b7-4e5b-9cca-521f76814fb2 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1011.113577] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-50f0b427-d578-4788-b2d0-b00d8ee3b319 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.120948] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for the task: (returnval){ [ 1011.120948] env[62000]: value = "task-882796" [ 1011.120948] env[62000]: _type = "Task" [ 1011.120948] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.130441] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882796, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.388898] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1011.394829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.493060] env[62000]: DEBUG oslo_concurrency.lockutils [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] Releasing lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.493417] env[62000]: DEBUG nova.compute.manager [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Received event network-vif-deleted-ad177e0f-5b23-4aa7-87ae-52f9de2e6650 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1011.493624] env[62000]: INFO nova.compute.manager [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Neutron deleted interface ad177e0f-5b23-4aa7-87ae-52f9de2e6650; detaching it from the instance and deleting it from the info cache [ 1011.493811] env[62000]: DEBUG nova.network.neutron [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.519904] env[62000]: DEBUG nova.scheduler.client.report [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1011.567375] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882794, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.570981] env[62000]: DEBUG nova.network.neutron [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Updating instance_info_cache with network_info: [{"id": "37a37223-a542-43e4-8527-64b1beecba38", "address": "fa:16:3e:10:2f:64", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37a37223-a5", "ovs_interfaceid": "37a37223-a542-43e4-8527-64b1beecba38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.574997] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.632652] env[62000]: DEBUG oslo_vmware.api [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Task: {'id': task-882796, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.38491} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.632908] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.633115] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1011.633540] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1011.633540] env[62000]: INFO nova.compute.manager [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1011.633755] env[62000]: DEBUG oslo.service.loopingcall [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.633945] env[62000]: DEBUG nova.compute.manager [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.634048] env[62000]: DEBUG nova.network.neutron [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1011.709588] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.735310] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.735571] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.735723] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.736753] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.736910] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.737045] env[62000]: DEBUG nova.virt.hardware [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.737891] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1d201e7-212e-40df-bd85-c974f29d4f78 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.746475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4222e49a-f247-4109-83a7-55263aa2334f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.996578] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aca9c9d5-69c2-4f53-a160-159162533236 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.008223] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26b6dea-9180-40dc-98c1-a23ddcac207f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.038462] env[62000]: DEBUG nova.compute.manager [req-f0a60457-0abe-4660-9333-4f724ec0d815 req-cada2c9c-0738-4cea-a8c9-342000d119c5 service nova] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Detach interface failed, port_id=ad177e0f-5b23-4aa7-87ae-52f9de2e6650, reason: Instance 9f864cad-f454-4dc3-b6d0-793d7adec0ae could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1012.065667] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882794, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643204} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.065935] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 9ccf07fa-90db-40b9-af5d-b47ee5773f9c/9ccf07fa-90db-40b9-af5d-b47ee5773f9c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1012.066175] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1012.066428] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6e560368-9645-43f7-9f11-a0a501de095e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.073425] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1012.073720] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Instance network_info: |[{"id": "37a37223-a542-43e4-8527-64b1beecba38", "address": "fa:16:3e:10:2f:64", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37a37223-a5", "ovs_interfaceid": "37a37223-a542-43e4-8527-64b1beecba38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1012.074055] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1012.074055] env[62000]: value = "task-882797" [ 1012.074055] env[62000]: _type = "Task" [ 1012.074055] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.074379] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:2f:64', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37a37223-a542-43e4-8527-64b1beecba38', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1012.087232] env[62000]: DEBUG oslo.service.loopingcall [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1012.088923] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1012.094445] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Received event network-vif-plugged-37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.094659] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Acquiring lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.094869] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.095050] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.095268] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] No waiting events found dispatching network-vif-plugged-37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1012.095450] env[62000]: WARNING nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Received unexpected event network-vif-plugged-37a37223-a542-43e4-8527-64b1beecba38 for instance with vm_state building and task_state spawning. [ 1012.095618] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Received event network-changed-37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1012.095774] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Refreshing instance network info cache due to event network-changed-37a37223-a542-43e4-8527-64b1beecba38. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1012.095957] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Acquiring lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.096112] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Acquired lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.096273] env[62000]: DEBUG nova.network.neutron [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Refreshing network info cache for port 37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1012.097390] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-73d0a398-4c8a-4459-830b-d9f9aab7e7dc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.118850] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882797, 'name': ExtendVirtualDisk_Task} progress is 50%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.120255] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1012.120255] env[62000]: value = "task-882798" [ 1012.120255] env[62000]: _type = "Task" [ 1012.120255] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.129249] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882798, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.341055] env[62000]: DEBUG nova.network.neutron [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.440118] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Successfully updated port: e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.541569] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.854s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.546597] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.614s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.548054] env[62000]: INFO nova.compute.claims [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1012.598790] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882797, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072996} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.599066] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1012.599821] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a95dffb9-c0d4-4180-b1c9-54a904784996 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.623421] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfiguring VM instance instance-00000062 to attach disk [datastore2] 9ccf07fa-90db-40b9-af5d-b47ee5773f9c/9ccf07fa-90db-40b9-af5d-b47ee5773f9c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1012.623697] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ed309199-3608-4104-8b21-bd1e092b6e27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.646671] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882798, 'name': CreateVM_Task, 'duration_secs': 0.400973} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.647685] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1012.648034] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1012.648034] env[62000]: value = "task-882799" [ 1012.648034] env[62000]: _type = "Task" [ 1012.648034] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.648800] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.648800] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.649134] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1012.649427] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4f56427-acfc-4ec4-9da5-68363fdd5d7a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.656810] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1012.656810] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52228554-aabe-c034-047b-db89de558433" [ 1012.656810] env[62000]: _type = "Task" [ 1012.656810] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.659985] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882799, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.667699] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52228554-aabe-c034-047b-db89de558433, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.835115] env[62000]: DEBUG nova.network.neutron [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Updated VIF entry in instance network info cache for port 37a37223-a542-43e4-8527-64b1beecba38. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1012.835567] env[62000]: DEBUG nova.network.neutron [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Updating instance_info_cache with network_info: [{"id": "37a37223-a542-43e4-8527-64b1beecba38", "address": "fa:16:3e:10:2f:64", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37a37223-a5", "ovs_interfaceid": "37a37223-a542-43e4-8527-64b1beecba38", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.847023] env[62000]: INFO nova.compute.manager [-] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Took 1.21 seconds to deallocate network for instance. [ 1012.945576] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.945743] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.945947] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1013.163050] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882799, 'name': ReconfigVM_Task, 'duration_secs': 0.292544} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.167064] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfigured VM instance instance-00000062 to attach disk [datastore2] 9ccf07fa-90db-40b9-af5d-b47ee5773f9c/9ccf07fa-90db-40b9-af5d-b47ee5773f9c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1013.167819] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d496e652-0e37-4a38-bf9f-8487a7810959 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.176585] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52228554-aabe-c034-047b-db89de558433, 'name': SearchDatastore_Task, 'duration_secs': 0.014528} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.178119] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.178420] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1013.178730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.178950] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.179167] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1013.179478] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1013.179478] env[62000]: value = "task-882800" [ 1013.179478] env[62000]: _type = "Task" [ 1013.179478] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.179667] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-feb0c818-774a-451d-a1a8-911234e5a3c0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.190611] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882800, 'name': Rename_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.191831] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1013.192041] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1013.192847] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da28e1fc-7088-4c2e-8f2b-3da812c6f59f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.198562] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1013.198562] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278fed9-2346-a702-c228-652b6320674a" [ 1013.198562] env[62000]: _type = "Task" [ 1013.198562] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.206987] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278fed9-2346-a702-c228-652b6320674a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.338794] env[62000]: DEBUG oslo_concurrency.lockutils [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] Releasing lock "refresh_cache-64ca26bd-dc8c-4f00-bfde-a24f8d650848" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.339178] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Received event network-vif-deleted-c1ca3821-1188-40f4-9056-a9d3b36b7a00 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1013.339283] env[62000]: INFO nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Neutron deleted interface c1ca3821-1188-40f4-9056-a9d3b36b7a00; detaching it from the instance and deleting it from the info cache [ 1013.339460] env[62000]: DEBUG nova.network.neutron [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.353118] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.483922] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1013.630309] env[62000]: DEBUG nova.network.neutron [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Updating instance_info_cache with network_info: [{"id": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "address": "fa:16:3e:a5:74:39", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape03a2c30-fa", "ovs_interfaceid": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.690487] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882800, 'name': Rename_Task, 'duration_secs': 0.146787} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.692737] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1013.693202] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-55389381-d121-40cd-b59c-54303b2c13a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.699817] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1013.699817] env[62000]: value = "task-882801" [ 1013.699817] env[62000]: _type = "Task" [ 1013.699817] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.713597] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5278fed9-2346-a702-c228-652b6320674a, 'name': SearchDatastore_Task, 'duration_secs': 0.018334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.717019] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882801, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.719341] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ec75401f-1ea0-487f-a6ad-8dfd9c8ac588 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.726096] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1013.726096] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523c8756-5d9d-1521-783f-0b4378cd9f9a" [ 1013.726096] env[62000]: _type = "Task" [ 1013.726096] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.736257] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523c8756-5d9d-1521-783f-0b4378cd9f9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.747087] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ae11630-d58d-4bf7-bfc2-8a0f1ab14226 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.754838] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4886ec0-3d4d-463e-82b9-805baba34445 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.787108] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a61300-2707-440b-9d84-bda8473c5c27 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.794797] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c78caff-82c7-499a-b7fc-34f9859b7ca3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.807973] env[62000]: DEBUG nova.compute.provider_tree [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1013.842510] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-02ada552-7822-4f45-bf70-9eda940af516 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.852414] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ce478e-ab55-4133-b72e-6ba6dcb1f2ec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.882381] env[62000]: DEBUG nova.compute.manager [req-88939ba1-eda7-4d07-bb69-ff706e638a6f req-19c7c36b-f58d-4cb7-b768-279d492e6260 service nova] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Detach interface failed, port_id=c1ca3821-1188-40f4-9056-a9d3b36b7a00, reason: Instance a2166191-09b7-4e5b-9cca-521f76814fb2 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1014.086955] env[62000]: INFO nova.compute.manager [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Swapping old allocation on dict_keys(['bcfccf20-49dd-4b91-819e-4373e67bf5ec']) held by migration 5793fe93-6df4-44ae-af50-d526a12184f6 for instance [ 1014.108412] env[62000]: DEBUG nova.compute.manager [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Received event network-vif-plugged-e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.108759] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Acquiring lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1014.108887] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.109021] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.109211] env[62000]: DEBUG nova.compute.manager [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] No waiting events found dispatching network-vif-plugged-e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1014.109418] env[62000]: WARNING nova.compute.manager [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Received unexpected event network-vif-plugged-e03a2c30-fa4a-4149-9e55-2631031aa8b0 for instance with vm_state building and task_state spawning. [ 1014.109588] env[62000]: DEBUG nova.compute.manager [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Received event network-changed-e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1014.109757] env[62000]: DEBUG nova.compute.manager [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Refreshing instance network info cache due to event network-changed-e03a2c30-fa4a-4149-9e55-2631031aa8b0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1014.109972] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Acquiring lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.111943] env[62000]: DEBUG nova.scheduler.client.report [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Overwriting current allocation {'allocations': {'bcfccf20-49dd-4b91-819e-4373e67bf5ec': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 129}}, 'project_id': '4d131f95ad3949d89cd6f36f6648d3f1', 'user_id': 'c3018dc2fa8d413aa464bf4e0c32341e', 'consumer_generation': 1} on consumer 3878579f-6435-4fe3-9f8c-8461d8ac57ee {{(pid=62000) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1014.132879] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.133229] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Instance network_info: |[{"id": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "address": "fa:16:3e:a5:74:39", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape03a2c30-fa", "ovs_interfaceid": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1014.133546] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Acquired lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.133733] env[62000]: DEBUG nova.network.neutron [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Refreshing network info cache for port e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1014.135134] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:74:39', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e03a2c30-fa4a-4149-9e55-2631031aa8b0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1014.144569] env[62000]: DEBUG oslo.service.loopingcall [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1014.147411] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1014.148701] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bed4d9cb-07a0-4bcd-99a4-7ee81b24646e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.172861] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1014.172861] env[62000]: value = "task-882802" [ 1014.172861] env[62000]: _type = "Task" [ 1014.172861] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.181913] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882802, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.213696] env[62000]: DEBUG oslo_vmware.api [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882801, 'name': PowerOnVM_Task, 'duration_secs': 0.469428} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.214010] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1014.214931] env[62000]: INFO nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Took 7.06 seconds to spawn the instance on the hypervisor. [ 1014.214931] env[62000]: DEBUG nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1014.215331] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347a4f6d-a7cd-4cc7-85ef-0284be6d344f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.225463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.225707] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.225890] env[62000]: DEBUG nova.network.neutron [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1014.242926] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]523c8756-5d9d-1521-783f-0b4378cd9f9a, 'name': SearchDatastore_Task, 'duration_secs': 0.012835} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.243434] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.243677] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 64ca26bd-dc8c-4f00-bfde-a24f8d650848/64ca26bd-dc8c-4f00-bfde-a24f8d650848.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1014.244853] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fea728ed-91ca-4ea9-8ea0-7c9ab73580f3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.252422] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1014.252422] env[62000]: value = "task-882803" [ 1014.252422] env[62000]: _type = "Task" [ 1014.252422] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.263694] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882803, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.311470] env[62000]: DEBUG nova.scheduler.client.report [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1014.415508] env[62000]: DEBUG nova.network.neutron [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Updated VIF entry in instance network info cache for port e03a2c30-fa4a-4149-9e55-2631031aa8b0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1014.415918] env[62000]: DEBUG nova.network.neutron [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Updating instance_info_cache with network_info: [{"id": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "address": "fa:16:3e:a5:74:39", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape03a2c30-fa", "ovs_interfaceid": "e03a2c30-fa4a-4149-9e55-2631031aa8b0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.684036] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882802, 'name': CreateVM_Task, 'duration_secs': 0.357531} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.684431] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1014.685063] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.685254] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.685598] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.686233] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d55940e5-17e4-4e87-9926-3d88f8544c79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.691876] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1014.691876] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529cf32e-8773-be24-b48a-542cb969e548" [ 1014.691876] env[62000]: _type = "Task" [ 1014.691876] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.700673] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529cf32e-8773-be24-b48a-542cb969e548, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.736601] env[62000]: INFO nova.compute.manager [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Took 21.91 seconds to build instance. [ 1014.764383] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882803, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.817180] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1014.817603] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1014.820394] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.426s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1014.820633] env[62000]: DEBUG nova.objects.instance [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lazy-loading 'resources' on Instance uuid 9f864cad-f454-4dc3-b6d0-793d7adec0ae {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.919013] env[62000]: DEBUG oslo_concurrency.lockutils [req-0e265cf2-6f70-43e8-b2e3-9597e5584e46 req-0acafb65-86de-4bdc-9430-a8b29ada2bee service nova] Releasing lock "refresh_cache-67f8274b-a0ad-419a-81fc-515b06ad41aa" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.019802] env[62000]: DEBUG nova.network.neutron [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [{"id": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "address": "fa:16:3e:16:ab:6a", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1eb9c5c5-8f", "ovs_interfaceid": "1eb9c5c5-8fe1-4652-986f-3b9abdd36485", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.203503] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529cf32e-8773-be24-b48a-542cb969e548, 'name': SearchDatastore_Task, 'duration_secs': 0.019814} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.203873] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.204128] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1015.204418] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.204601] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.204888] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1015.205118] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fbc9430-a2c9-4475-99f0-0f3f9ca68f0b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.221147] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1015.221347] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1015.222159] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71e36d8f-34eb-439e-b8a5-5aa1c0dd90c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.228620] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1015.228620] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527480b6-bb3c-b85a-0c27-4c2aa64ddcd1" [ 1015.228620] env[62000]: _type = "Task" [ 1015.228620] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.237299] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527480b6-bb3c-b85a-0c27-4c2aa64ddcd1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.238763] env[62000]: DEBUG oslo_concurrency.lockutils [None req-fb7e2386-9d77-408e-9f8b-0abfa8efcedf tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.419s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.262361] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882803, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.603905} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.262681] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 64ca26bd-dc8c-4f00-bfde-a24f8d650848/64ca26bd-dc8c-4f00-bfde-a24f8d650848.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1015.262910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1015.263233] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25b9aadf-ed5e-4a4a-b623-21f48a573a13 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.271161] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1015.271161] env[62000]: value = "task-882804" [ 1015.271161] env[62000]: _type = "Task" [ 1015.271161] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.280020] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882804, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.323834] env[62000]: DEBUG nova.compute.utils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1015.328482] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1015.328482] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1015.403940] env[62000]: DEBUG nova.policy [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '79730e91610c4c598019b122b4f41418', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6106a70abdad4111977eebbd82434337', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1015.488346] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4f59ef-27a0-4bf7-ab53-923c83e2750b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.496985] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b0d11c-aaae-44b5-82c6-72344333cb53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.529176] env[62000]: DEBUG oslo_concurrency.lockutils [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-3878579f-6435-4fe3-9f8c-8461d8ac57ee" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.530279] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f892065-214f-40c7-b1cc-031eae187ded {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.533574] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5417c1a-f903-40d2-88e3-999674b9cb21 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.543365] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ea605f-6934-496b-8d3b-f75e10b85d00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.549424] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82d08907-9566-468b-acb8-39606edbfb1c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.564051] env[62000]: DEBUG nova.compute.provider_tree [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1015.743747] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]527480b6-bb3c-b85a-0c27-4c2aa64ddcd1, 'name': SearchDatastore_Task, 'duration_secs': 0.052585} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.744863] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b63b500f-966c-41cf-bd65-bca209916900 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.752194] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1015.752194] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b9204c-9f8b-e9f8-2fb6-ba849bb48aa9" [ 1015.752194] env[62000]: _type = "Task" [ 1015.752194] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.760032] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b9204c-9f8b-e9f8-2fb6-ba849bb48aa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.786531] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882804, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074907} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.786813] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1015.787603] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-755d5bea-2b93-4060-900d-df1b29ae60c2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.812345] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Reconfiguring VM instance instance-00000063 to attach disk [datastore2] 64ca26bd-dc8c-4f00-bfde-a24f8d650848/64ca26bd-dc8c-4f00-bfde-a24f8d650848.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.812658] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a753003c-8919-4938-ac3f-a87d0d377d48 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.832656] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1015.838743] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1015.838743] env[62000]: value = "task-882805" [ 1015.838743] env[62000]: _type = "Task" [ 1015.838743] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.851425] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882805, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.955186] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Successfully created port: 9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1016.072520] env[62000]: DEBUG nova.scheduler.client.report [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1016.264056] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b9204c-9f8b-e9f8-2fb6-ba849bb48aa9, 'name': SearchDatastore_Task, 'duration_secs': 0.072121} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.264368] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.264688] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 67f8274b-a0ad-419a-81fc-515b06ad41aa/67f8274b-a0ad-419a-81fc-515b06ad41aa.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1016.265010] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-43b8ed5f-1bcb-42e6-a712-b0b63df544f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.273958] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1016.273958] env[62000]: value = "task-882806" [ 1016.273958] env[62000]: _type = "Task" [ 1016.273958] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.282484] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.343117] env[62000]: INFO nova.virt.block_device [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Booting with volume d6f79cc7-55cc-4d38-9301-08c8d382294f at /dev/sda [ 1016.355729] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882805, 'name': ReconfigVM_Task, 'duration_secs': 0.488838} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.356219] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Reconfigured VM instance instance-00000063 to attach disk [datastore2] 64ca26bd-dc8c-4f00-bfde-a24f8d650848/64ca26bd-dc8c-4f00-bfde-a24f8d650848.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.356882] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5254218f-2155-462b-b6a4-600577c466d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.365139] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1016.365139] env[62000]: value = "task-882807" [ 1016.365139] env[62000]: _type = "Task" [ 1016.365139] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.373852] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882807, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.381933] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-634a811e-614d-491a-ae9a-5f429ff2ccca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.391806] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856bf512-56d6-4697-8405-faca5dec2cde {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.432464] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-da4529a2-ba97-47e4-a1cf-7675dcd7f90c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.442898] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38131a71-dcb1-4679-97f6-0c217ffb4274 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.480046] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ce4e57-9ae4-4a5c-b140-7c8908242dee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.488472] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-776f633b-b1c4-4e65-9d14-6cd1226f5a47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.508028] env[62000]: DEBUG nova.virt.block_device [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating existing volume attachment record: ac743b5f-de32-45c3-ac43-579d33e281a5 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1016.579715] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1016.582751] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.008s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1016.584719] env[62000]: INFO nova.compute.claims [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.627512] env[62000]: INFO nova.scheduler.client.report [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Deleted allocations for instance 9f864cad-f454-4dc3-b6d0-793d7adec0ae [ 1016.637645] env[62000]: DEBUG nova.compute.manager [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Received event network-changed-5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1016.637942] env[62000]: DEBUG nova.compute.manager [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Refreshing instance network info cache due to event network-changed-5f0f0c25-4143-4468-b1e5-48897bd334ac. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1016.638217] env[62000]: DEBUG oslo_concurrency.lockutils [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] Acquiring lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.638519] env[62000]: DEBUG oslo_concurrency.lockutils [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] Acquired lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.638646] env[62000]: DEBUG nova.network.neutron [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Refreshing network info cache for port 5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1016.662696] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1016.662992] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f3d3c33-cccf-4078-8872-bf0bf883db29 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.674596] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1016.674596] env[62000]: value = "task-882808" [ 1016.674596] env[62000]: _type = "Task" [ 1016.674596] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.686920] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.787255] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882806, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.877776] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882807, 'name': Rename_Task, 'duration_secs': 0.161542} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.878197] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1016.878602] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6f4ea9b-156f-407e-ac79-ef43fb6ad6d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.886858] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1016.886858] env[62000]: value = "task-882809" [ 1016.886858] env[62000]: _type = "Task" [ 1016.886858] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.895689] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882809, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.144637] env[62000]: DEBUG oslo_concurrency.lockutils [None req-571fec2d-bf23-43c0-9738-2478a09010ed tempest-ServerMetadataTestJSON-1036926282 tempest-ServerMetadataTestJSON-1036926282-project-member] Lock "9f864cad-f454-4dc3-b6d0-793d7adec0ae" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.802s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.188340] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882808, 'name': PowerOffVM_Task, 'duration_secs': 0.240716} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.191617] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1017.191617] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1017.191987] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1017.192281] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1017.193035] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1017.193290] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1017.193617] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1017.194036] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1017.194521] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1017.194944] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1017.195890] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1017.196206] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1017.205125] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25894f92-5c0c-4347-a31c-15419897bc07 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.223365] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1017.223365] env[62000]: value = "task-882810" [ 1017.223365] env[62000]: _type = "Task" [ 1017.223365] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.232420] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882810, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.287566] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.534834} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.287755] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 67f8274b-a0ad-419a-81fc-515b06ad41aa/67f8274b-a0ad-419a-81fc-515b06ad41aa.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1017.287863] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.288110] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d67a8304-1f4a-4e46-86a5-a5157e972826 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.295719] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1017.295719] env[62000]: value = "task-882811" [ 1017.295719] env[62000]: _type = "Task" [ 1017.295719] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.307368] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882811, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.396258] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882809, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.432318] env[62000]: DEBUG nova.network.neutron [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updated VIF entry in instance network info cache for port 5f0f0c25-4143-4468-b1e5-48897bd334ac. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1017.432714] env[62000]: DEBUG nova.network.neutron [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating instance_info_cache with network_info: [{"id": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "address": "fa:16:3e:40:72:07", "network": {"id": "dce2389b-f58a-4b60-88c4-411a27993ef6", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-918286612-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.244", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6742b3d48eb945fe8a11d0ee651abd1a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f5f49a7d-c6e5-404f-b71a-91d8c070cd18", "external-id": "nsx-vlan-transportzone-120", "segmentation_id": 120, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5f0f0c25-41", "ovs_interfaceid": "5f0f0c25-4143-4468-b1e5-48897bd334ac", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.735886] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882810, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.756647] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-141a1af8-3ea1-44df-9e6d-08d69359c99a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.765500] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-994617b0-a00a-4fe4-b41a-e87d8c30d472 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.800955] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f780d262-dec9-4f9e-9a51-77f497c89509 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.815081] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5ec173c-3717-4aa3-8d70-b743062ee173 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.818141] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882811, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080252} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.818423] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.819559] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb44f4c-0851-4bc2-8f53-e9235adf404a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.833178] env[62000]: DEBUG nova.compute.provider_tree [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1017.856066] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 67f8274b-a0ad-419a-81fc-515b06ad41aa/67f8274b-a0ad-419a-81fc-515b06ad41aa.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.857186] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d93e8860-f7d9-4cfe-9258-0fdc70df529c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.878816] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1017.878816] env[62000]: value = "task-882812" [ 1017.878816] env[62000]: _type = "Task" [ 1017.878816] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.887750] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882812, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.896464] env[62000]: DEBUG oslo_vmware.api [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882809, 'name': PowerOnVM_Task, 'duration_secs': 1.003951} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.896718] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1017.896916] env[62000]: INFO nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Took 8.47 seconds to spawn the instance on the hypervisor. [ 1017.897111] env[62000]: DEBUG nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1017.897887] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4017638d-3762-42ab-9ec3-fa10fccf067b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.935910] env[62000]: DEBUG oslo_concurrency.lockutils [req-bc405797-e097-42c5-9655-3f104023cb2f req-4c612424-9482-44ac-a3f1-2b791e21b64e service nova] Releasing lock "refresh_cache-9ccf07fa-90db-40b9-af5d-b47ee5773f9c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.990118] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Successfully updated port: 9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1018.234406] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882810, 'name': ReconfigVM_Task, 'duration_secs': 0.516731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.235339] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92144f79-ea50-4572-a22c-4a424a4d3a0e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.256236] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.256633] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.256745] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.256890] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.257053] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.257213] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.257480] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.257677] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.257858] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.258038] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.258222] env[62000]: DEBUG nova.virt.hardware [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.259174] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d04f5c4-ba4f-48c4-aebe-53eca945e7a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.265315] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1018.265315] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f379a-9b53-9b26-d8ba-af8e4a31d5f1" [ 1018.265315] env[62000]: _type = "Task" [ 1018.265315] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.273826] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f379a-9b53-9b26-d8ba-af8e4a31d5f1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.336439] env[62000]: DEBUG nova.scheduler.client.report [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1018.391780] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882812, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.416719] env[62000]: INFO nova.compute.manager [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Took 25.12 seconds to build instance. [ 1018.492811] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.492959] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.493282] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1018.613737] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1018.614295] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1018.615174] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1018.615174] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1018.615174] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1018.615174] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1018.615174] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1018.615562] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1018.615562] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1018.615659] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1018.615806] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1018.615981] env[62000]: DEBUG nova.virt.hardware [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1018.616862] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f82c4f-8ab3-49bd-b99a-75d85ff9ba61 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.626370] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac84a838-8976-4c1a-81d1-854a670bea4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.661094] env[62000]: DEBUG nova.compute.manager [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Received event network-vif-plugged-9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.661315] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.661527] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.661698] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.661867] env[62000]: DEBUG nova.compute.manager [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] No waiting events found dispatching network-vif-plugged-9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1018.662148] env[62000]: WARNING nova.compute.manager [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Received unexpected event network-vif-plugged-9f791142-c29f-4dda-b815-14d5e18a8f8c for instance with vm_state building and task_state spawning. [ 1018.662447] env[62000]: DEBUG nova.compute.manager [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Received event network-changed-9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1018.662628] env[62000]: DEBUG nova.compute.manager [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Refreshing instance network info cache due to event network-changed-9f791142-c29f-4dda-b815-14d5e18a8f8c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1018.662806] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Acquiring lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.775964] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520f379a-9b53-9b26-d8ba-af8e4a31d5f1, 'name': SearchDatastore_Task, 'duration_secs': 0.00918} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.781389] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1018.781693] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6c7da688-874c-4f3a-a4b3-e923a31dc750 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.803466] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1018.803466] env[62000]: value = "task-882813" [ 1018.803466] env[62000]: _type = "Task" [ 1018.803466] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.810769] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882813, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.841352] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.259s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1018.841997] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1018.845277] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.492s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.845446] env[62000]: DEBUG nova.objects.instance [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lazy-loading 'resources' on Instance uuid a2166191-09b7-4e5b-9cca-521f76814fb2 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.890906] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882812, 'name': ReconfigVM_Task, 'duration_secs': 0.541074} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.891132] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 67f8274b-a0ad-419a-81fc-515b06ad41aa/67f8274b-a0ad-419a-81fc-515b06ad41aa.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1018.891780] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93d3a9b6-a2dd-45fc-b462-b9faaae2ac7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.899716] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1018.899716] env[62000]: value = "task-882814" [ 1018.899716] env[62000]: _type = "Task" [ 1018.899716] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.908347] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882814, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.917941] env[62000]: DEBUG oslo_concurrency.lockutils [None req-842d0896-67b6-487a-b4c3-34bb97edafa2 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.635s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.028523] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1019.049316] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.049422] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.049582] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.049760] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.051859] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.052763] env[62000]: INFO nova.compute.manager [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Terminating instance [ 1019.056238] env[62000]: DEBUG nova.compute.manager [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1019.056446] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1019.057330] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b0f4cc9-be39-4038-9e7e-c3c45c4d114b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.066150] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1019.066537] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b147a712-7e2d-45e8-bdda-8f61bf20720b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.074270] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1019.074270] env[62000]: value = "task-882815" [ 1019.074270] env[62000]: _type = "Task" [ 1019.074270] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.086962] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.238016] env[62000]: DEBUG nova.network.neutron [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [{"id": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "address": "fa:16:3e:ee:fd:8d", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f791142-c2", "ovs_interfaceid": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.313088] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882813, 'name': ReconfigVM_Task, 'duration_secs': 0.257706} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.313401] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1019.314198] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9559ae77-067a-4b85-944d-9c998b2a86e6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.342263] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1019.342603] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b784a38-50ea-48c9-9456-f5b34fa2cd17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.357965] env[62000]: DEBUG nova.compute.utils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1019.363192] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1019.363383] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1019.374512] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1019.374512] env[62000]: value = "task-882816" [ 1019.374512] env[62000]: _type = "Task" [ 1019.374512] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.385807] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882816, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.413770] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882814, 'name': Rename_Task, 'duration_secs': 0.187134} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.415372] env[62000]: DEBUG nova.policy [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfe15ae102584204b0816ce6e36afdfd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cac5f0a5704d434082131155e107d190', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1019.417028] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.417323] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-961abef3-a227-49ca-9b43-b4c1407caf23 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.425445] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1019.425445] env[62000]: value = "task-882817" [ 1019.425445] env[62000]: _type = "Task" [ 1019.425445] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.434426] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882817, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.541243] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96ace6e1-a9c9-46a8-9ec8-3ccbe65182f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.552407] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21cb45b0-aeac-4faf-a815-c5465f315b8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.588706] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946f7502-469f-4383-bd82-afbedb88391e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.600369] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ac3bb4-bc2e-4f4d-bd68-f1294358f20a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.604858] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882815, 'name': PowerOffVM_Task, 'duration_secs': 0.250631} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.605271] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1019.605937] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1019.606288] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c43a2955-f6b2-4507-8018-2c7b935c0165 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.621197] env[62000]: DEBUG nova.compute.provider_tree [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.677801] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1019.678062] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1019.678329] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore2] 64ca26bd-dc8c-4f00-bfde-a24f8d650848 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1019.678612] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-27e33880-e224-4fd5-ad3e-1d3fb0143587 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.685625] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1019.685625] env[62000]: value = "task-882819" [ 1019.685625] env[62000]: _type = "Task" [ 1019.685625] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.694847] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.740944] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.741415] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance network_info: |[{"id": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "address": "fa:16:3e:ee:fd:8d", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f791142-c2", "ovs_interfaceid": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1019.741797] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Acquired lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.742120] env[62000]: DEBUG nova.network.neutron [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Refreshing network info cache for port 9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1019.743543] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ee:fd:8d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e044cfd4-1b0d-4d88-b1bd-604025731d3f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9f791142-c29f-4dda-b815-14d5e18a8f8c', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1019.752185] env[62000]: DEBUG oslo.service.loopingcall [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1019.755759] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1019.756502] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-737defef-60a3-43e7-b3ee-05b9d744f948 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.783775] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1019.783775] env[62000]: value = "task-882820" [ 1019.783775] env[62000]: _type = "Task" [ 1019.783775] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.793915] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882820, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.863715] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1019.867768] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Successfully created port: 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1019.886692] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882816, 'name': ReconfigVM_Task, 'duration_secs': 0.434408} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.887228] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee/3878579f-6435-4fe3-9f8c-8461d8ac57ee.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.888548] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d504c70c-6707-4a07-8c2c-f24586aaaf16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.918927] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd2b4f0-4f42-4904-bc1c-8a9a55380d38 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.954634] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44c441d6-1f6b-487c-9f5d-272530a4dc32 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.961868] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882817, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.983461] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f82d69-c4d5-4641-95e5-0e1ae2952f98 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.992185] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1019.992530] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b49f9ff3-e1d4-4acb-9ae6-663d212dffaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.004124] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1020.004124] env[62000]: value = "task-882821" [ 1020.004124] env[62000]: _type = "Task" [ 1020.004124] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.010324] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882821, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.126126] env[62000]: DEBUG nova.scheduler.client.report [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1020.192463] env[62000]: DEBUG nova.network.neutron [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updated VIF entry in instance network info cache for port 9f791142-c29f-4dda-b815-14d5e18a8f8c. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1020.193059] env[62000]: DEBUG nova.network.neutron [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [{"id": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "address": "fa:16:3e:ee:fd:8d", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f791142-c2", "ovs_interfaceid": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.199984] env[62000]: DEBUG oslo_vmware.api [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203095} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.200239] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1020.200425] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1020.200602] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1020.200776] env[62000]: INFO nova.compute.manager [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1020.201033] env[62000]: DEBUG oslo.service.loopingcall [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1020.201375] env[62000]: DEBUG nova.compute.manager [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1020.201375] env[62000]: DEBUG nova.network.neutron [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1020.294097] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882820, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.436480] env[62000]: DEBUG oslo_vmware.api [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882817, 'name': PowerOnVM_Task, 'duration_secs': 0.576266} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.436896] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1020.437017] env[62000]: INFO nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1020.437193] env[62000]: DEBUG nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1020.437962] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5967f85-ba01-4335-b72c-f8ba6fe381ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.512721] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882821, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.632072] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.787s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.659807] env[62000]: INFO nova.scheduler.client.report [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Deleted allocations for instance a2166191-09b7-4e5b-9cca-521f76814fb2 [ 1020.690064] env[62000]: DEBUG nova.compute.manager [req-a9d61910-9b81-4bd5-932a-c7acbd901626 req-da51310d-7f83-4756-bc7f-9c2a8fe76a06 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Received event network-vif-deleted-37a37223-a542-43e4-8527-64b1beecba38 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1020.690296] env[62000]: INFO nova.compute.manager [req-a9d61910-9b81-4bd5-932a-c7acbd901626 req-da51310d-7f83-4756-bc7f-9c2a8fe76a06 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Neutron deleted interface 37a37223-a542-43e4-8527-64b1beecba38; detaching it from the instance and deleting it from the info cache [ 1020.690486] env[62000]: DEBUG nova.network.neutron [req-a9d61910-9b81-4bd5-932a-c7acbd901626 req-da51310d-7f83-4756-bc7f-9c2a8fe76a06 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.695974] env[62000]: DEBUG oslo_concurrency.lockutils [req-7c31fc41-5bfd-4397-b03a-a89385461c98 req-47a67e71-c865-465e-bd60-c5eec5a45f20 service nova] Releasing lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.795632] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882820, 'name': CreateVM_Task, 'duration_secs': 0.517167} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.795813] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1020.796508] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'attachment_id': 'ac743b5f-de32-45c3-ac43-579d33e281a5', 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201633', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'name': 'volume-d6f79cc7-55cc-4d38-9301-08c8d382294f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'serial': 'd6f79cc7-55cc-4d38-9301-08c8d382294f'}, 'mount_device': '/dev/sda', 'device_type': None, 'disk_bus': None, 'delete_on_termination': True, 'boot_index': 0, 'guest_format': None, 'volume_type': None}], 'swap': None} {{(pid=62000) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1020.796724] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Root volume attach. Driver type: vmdk {{(pid=62000) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1020.797574] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-729c18d5-e0b6-4de2-b969-d8d723172fe8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.805320] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c49c243c-53d7-463a-95e1-1d6da90ef4b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.811231] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba11f2e-e820-40b5-934f-c7f3ddc817b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.816943] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-9df283c4-3c86-4472-822b-440067b5eff1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.823726] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1020.823726] env[62000]: value = "task-882822" [ 1020.823726] env[62000]: _type = "Task" [ 1020.823726] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.833561] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.881581] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1020.918986] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1020.919240] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1020.919400] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1020.919564] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1020.919717] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1020.919868] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1020.920132] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1020.920310] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1020.920489] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1020.920658] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1020.920833] env[62000]: DEBUG nova.virt.hardware [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1020.921724] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a925197c-35e6-4983-a7af-e167738b8d86 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.930779] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056d0172-4e2d-4b27-9475-c935e3eb0618 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.957923] env[62000]: INFO nova.compute.manager [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Took 23.95 seconds to build instance. [ 1021.011866] env[62000]: DEBUG oslo_vmware.api [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882821, 'name': PowerOnVM_Task, 'duration_secs': 0.616678} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.012312] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1021.065629] env[62000]: DEBUG nova.network.neutron [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1021.167865] env[62000]: DEBUG oslo_concurrency.lockutils [None req-047bf137-0bda-4699-92f0-07009ba14589 tempest-ServerAddressesTestJSON-1636557993 tempest-ServerAddressesTestJSON-1636557993-project-member] Lock "a2166191-09b7-4e5b-9cca-521f76814fb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.668s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.193402] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-097249b2-be80-4ce2-b198-389beba5f12a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.204581] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a42e2e4-a410-4f45-8699-430140bf2226 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.238601] env[62000]: DEBUG nova.compute.manager [req-a9d61910-9b81-4bd5-932a-c7acbd901626 req-da51310d-7f83-4756-bc7f-9c2a8fe76a06 service nova] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Detach interface failed, port_id=37a37223-a542-43e4-8527-64b1beecba38, reason: Instance 64ca26bd-dc8c-4f00-bfde-a24f8d650848 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1021.337191] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 35%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.460746] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a0156687-105a-4eb8-a2df-70bbe6e89a9c tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.462s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.483828] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec93fff5-a4df-4283-b867-777bce5fc464 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.493032] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Suspending the VM {{(pid=62000) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1163}} [ 1021.493032] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-1f1fedcd-d14d-4bb6-9dca-31aca826b77b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.501054] env[62000]: DEBUG oslo_vmware.api [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1021.501054] env[62000]: value = "task-882823" [ 1021.501054] env[62000]: _type = "Task" [ 1021.501054] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.513227] env[62000]: DEBUG oslo_vmware.api [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882823, 'name': SuspendVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.568199] env[62000]: INFO nova.compute.manager [-] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Took 1.37 seconds to deallocate network for instance. [ 1021.662212] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Successfully updated port: 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1021.841724] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 53%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.015699] env[62000]: DEBUG oslo_vmware.api [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882823, 'name': SuspendVM_Task} progress is 54%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.066295] env[62000]: INFO nova.compute.manager [None req-02a4015c-7850-4d35-a9c3-0b637b4cb6a3 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance to original state: 'active' [ 1022.081189] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.081489] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.081715] env[62000]: DEBUG nova.objects.instance [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid 64ca26bd-dc8c-4f00-bfde-a24f8d650848 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1022.164812] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.165013] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1022.165194] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1022.337537] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 65%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.513562] env[62000]: DEBUG oslo_vmware.api [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882823, 'name': SuspendVM_Task, 'duration_secs': 0.953034} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.513921] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Suspended the VM {{(pid=62000) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1167}} [ 1022.514173] env[62000]: DEBUG nova.compute.manager [None req-ba9ed0fa-8bf8-4be1-b062-4d7ec2173643 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.515064] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8401b401-3b88-4446-8800-ef737defbd8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.744396] env[62000]: DEBUG nova.compute.manager [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.744660] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.744874] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.745109] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.745329] env[62000]: DEBUG nova.compute.manager [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] No waiting events found dispatching network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1022.745548] env[62000]: WARNING nova.compute.manager [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received unexpected event network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 for instance with vm_state building and task_state spawning. [ 1022.745725] env[62000]: DEBUG nova.compute.manager [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1022.745885] env[62000]: DEBUG nova.compute.manager [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing instance network info cache due to event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1022.746085] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1022.748447] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1022.773650] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82c1d762-ff42-4aad-a843-209ac4040a90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.789346] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e80db19-0862-42eb-8de5-2adc0db250aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.833336] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe4986e3-fee7-4ebf-af21-66477606ade9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.845246] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 78%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.846722] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b288252-037f-40b7-a2d0-bc912c6887e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.866259] env[62000]: DEBUG nova.compute.provider_tree [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1023.036744] env[62000]: DEBUG nova.network.neutron [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.344883] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 92%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.369933] env[62000]: DEBUG nova.scheduler.client.report [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.536708] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.537076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.537430] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.537622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.537837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.540009] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1023.540356] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance network_info: |[{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1023.541081] env[62000]: INFO nova.compute.manager [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Terminating instance [ 1023.542624] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1023.542854] env[62000]: DEBUG nova.network.neutron [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1023.544027] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:1b:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ec45f71-6b9c-4a0a-bd0f-f73708c92053', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1023.557278] env[62000]: DEBUG oslo.service.loopingcall [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.559230] env[62000]: DEBUG nova.compute.manager [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.559546] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1023.561725] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1023.561919] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4571000b-43fe-4749-987d-984c1a905124 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.564039] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c048a4f2-6b65-40cf-9535-1a53a42a17cb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.586662] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1023.586662] env[62000]: value = "task-882825" [ 1023.586662] env[62000]: _type = "Task" [ 1023.586662] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.588103] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1023.588103] env[62000]: value = "task-882824" [ 1023.588103] env[62000]: _type = "Task" [ 1023.588103] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.599835] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882824, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.603078] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882825, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.841205] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.875320] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.794s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.902816] env[62000]: INFO nova.scheduler.client.report [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance 64ca26bd-dc8c-4f00-bfde-a24f8d650848 [ 1023.994523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.994523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.994523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.994523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.994523] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.996623] env[62000]: INFO nova.compute.manager [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Terminating instance [ 1023.998572] env[62000]: DEBUG nova.compute.manager [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1023.998779] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1023.999671] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa301785-3ee5-4155-8877-ee1f86c9bf76 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.008352] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1024.008627] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c4c41a9-4c93-4b51-883f-a5e378f5703f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.079628] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1024.079915] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1024.080099] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore1] 67f8274b-a0ad-419a-81fc-515b06ad41aa {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1024.080395] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-964794e4-f351-4e66-9cc5-2243d0adf797 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.088624] env[62000]: DEBUG oslo_vmware.api [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1024.088624] env[62000]: value = "task-882827" [ 1024.088624] env[62000]: _type = "Task" [ 1024.088624] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.102498] env[62000]: DEBUG oslo_vmware.api [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882827, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.109186] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882825, 'name': CreateVM_Task, 'duration_secs': 0.387955} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.109186] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882824, 'name': PowerOffVM_Task, 'duration_secs': 0.217574} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.111265] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1024.111595] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1024.111808] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1024.112070] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201628', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'name': 'volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '3878579f-6435-4fe3-9f8c-8461d8ac57ee', 'attached_at': '2024-09-16T07:01:15.000000', 'detached_at': '', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'serial': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1024.112805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.113050] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.113489] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1024.114067] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb47d6df-bd53-47e1-b5ae-18fe180e9981 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.116814] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b180deef-0ac3-47b5-b132-a06eb045f056 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.144504] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d571865-d323-4b45-9c7c-61ef580b96ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.151280] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1024.151280] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ac5a69-3e62-d43a-ab15-4379ee528fff" [ 1024.151280] env[62000]: _type = "Task" [ 1024.151280] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.159989] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdd77edc-6e5d-4680-8344-6807ba58afa8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.166679] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ac5a69-3e62-d43a-ab15-4379ee528fff, 'name': SearchDatastore_Task, 'duration_secs': 0.026965} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.167442] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.167717] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1024.167982] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.168173] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.168382] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1024.168652] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3ad3c6b4-612a-4064-8780-68d8036bd82f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.192339] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d08bea-21b6-4b1e-8512-862513037c14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.209929] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] The volume has not been displaced from its original location: [datastore2] volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7/volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1024.215607] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfiguring VM instance instance-00000056 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1024.219839] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cbd62b4-d42e-4d52-b6de-6ba980931005 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.233839] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1024.234140] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1024.236201] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4e4a27b7-4b5c-46fc-b8a0-8060ab7c68b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.242840] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1024.242840] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fa0a0b-e467-2d99-eec5-8fb3497ab807" [ 1024.242840] env[62000]: _type = "Task" [ 1024.242840] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.248112] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1024.248112] env[62000]: value = "task-882828" [ 1024.248112] env[62000]: _type = "Task" [ 1024.248112] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.254600] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fa0a0b-e467-2d99-eec5-8fb3497ab807, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.260637] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882828, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.340427] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task} progress is 98%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.410437] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d73ebdf4-9f04-42fc-8c40-acef6325fc64 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "64ca26bd-dc8c-4f00-bfde-a24f8d650848" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.361s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.418703] env[62000]: DEBUG nova.network.neutron [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updated VIF entry in instance network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1024.419209] env[62000]: DEBUG nova.network.neutron [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.603210] env[62000]: DEBUG oslo_vmware.api [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882827, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.342348} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.603506] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.603694] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1024.603864] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1024.604044] env[62000]: INFO nova.compute.manager [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1024.604686] env[62000]: DEBUG oslo.service.loopingcall [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1024.604686] env[62000]: DEBUG nova.compute.manager [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1024.604686] env[62000]: DEBUG nova.network.neutron [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1024.755914] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52fa0a0b-e467-2d99-eec5-8fb3497ab807, 'name': SearchDatastore_Task, 'duration_secs': 0.015955} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.757031] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d9fd08e1-2d77-4328-8b5b-1f6116d4789f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.762081] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882828, 'name': ReconfigVM_Task, 'duration_secs': 0.254056} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.762687] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Reconfigured VM instance instance-00000056 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1024.768463] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e7445401-cc9b-43c7-8f61-321e814663b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.778720] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1024.778720] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b21e23-d71f-03a4-e038-c2e803c8ac2a" [ 1024.778720] env[62000]: _type = "Task" [ 1024.778720] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.785452] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1024.785452] env[62000]: value = "task-882829" [ 1024.785452] env[62000]: _type = "Task" [ 1024.785452] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.788896] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b21e23-d71f-03a4-e038-c2e803c8ac2a, 'name': SearchDatastore_Task, 'duration_secs': 0.010887} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.791873] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.792225] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1024.792473] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d732274-1ea5-4c08-ae70-97ae0243d0ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.802067] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882829, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.802736] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1024.802736] env[62000]: value = "task-882830" [ 1024.802736] env[62000]: _type = "Task" [ 1024.802736] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.812280] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882830, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.840838] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882822, 'name': RelocateVM_Task, 'duration_secs': 3.647446} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.841214] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1024.841457] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201633', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'name': 'volume-d6f79cc7-55cc-4d38-9301-08c8d382294f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'serial': 'd6f79cc7-55cc-4d38-9301-08c8d382294f'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1024.842281] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d5d9fc-5951-4cdc-9c03-352ff2ef7872 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.860855] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-122a5e3b-5fc0-43ff-92ee-614d7fc03b9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.885719] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Reconfiguring VM instance instance-00000065 to attach disk [datastore1] volume-d6f79cc7-55cc-4d38-9301-08c8d382294f/volume-d6f79cc7-55cc-4d38-9301-08c8d382294f.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1024.888295] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b0813a0-3bd2-41b6-b9b2-69b1eb513d90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.909626] env[62000]: DEBUG nova.compute.manager [req-f5ce4dbf-d29b-4d22-a55c-2840e002ecb7 req-93d0d8b3-f03a-4209-9ebf-1d19e41d883f service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Received event network-vif-deleted-e03a2c30-fa4a-4149-9e55-2631031aa8b0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1024.909765] env[62000]: INFO nova.compute.manager [req-f5ce4dbf-d29b-4d22-a55c-2840e002ecb7 req-93d0d8b3-f03a-4209-9ebf-1d19e41d883f service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Neutron deleted interface e03a2c30-fa4a-4149-9e55-2631031aa8b0; detaching it from the instance and deleting it from the info cache [ 1024.910136] env[62000]: DEBUG nova.network.neutron [req-f5ce4dbf-d29b-4d22-a55c-2840e002ecb7 req-93d0d8b3-f03a-4209-9ebf-1d19e41d883f service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.918700] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1024.918700] env[62000]: value = "task-882831" [ 1024.918700] env[62000]: _type = "Task" [ 1024.918700] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1024.925257] env[62000]: DEBUG oslo_concurrency.lockutils [req-aa8e40d4-44e4-4146-b895-5f34b1b3cce9 req-66d52034-4fe1-47aa-8033-9231dd6c93ce service nova] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1024.931915] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882831, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.280752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.281098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.303379] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882829, 'name': ReconfigVM_Task, 'duration_secs': 0.175324} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.303920] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201628', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'name': 'volume-f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': '3878579f-6435-4fe3-9f8c-8461d8ac57ee', 'attached_at': '2024-09-16T07:01:15.000000', 'detached_at': '', 'volume_id': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7', 'serial': 'f21e6dba-4892-4d3a-9758-294a3d9d63f7'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1025.304283] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1025.308593] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f26cf5f-efc4-48ea-be76-644ef8ea0f17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.319281] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1025.322949] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d1730d9b-6a95-41de-a40a-71c2f885ef56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.324947] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882830, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.389391] env[62000]: DEBUG nova.network.neutron [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.392164] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1025.392448] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1025.392641] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleting the datastore file [datastore1] 3878579f-6435-4fe3-9f8c-8461d8ac57ee {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1025.393190] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3a7fc8b0-3da6-48c4-977e-fac1ad3278d6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.400967] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1025.400967] env[62000]: value = "task-882833" [ 1025.400967] env[62000]: _type = "Task" [ 1025.400967] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.410025] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882833, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.412970] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e915509e-5e21-44af-9f9b-5135dad65fa1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.426144] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0540b684-9107-440d-b5cf-dad349e3fc44 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.440036] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882831, 'name': ReconfigVM_Task, 'duration_secs': 0.434166} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.440295] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Reconfigured VM instance instance-00000065 to attach disk [datastore1] volume-d6f79cc7-55cc-4d38-9301-08c8d382294f/volume-d6f79cc7-55cc-4d38-9301-08c8d382294f.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1025.445008] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70525e2a-5a2a-4de0-8b2f-1d0652132c9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.463204] env[62000]: DEBUG nova.compute.manager [req-f5ce4dbf-d29b-4d22-a55c-2840e002ecb7 req-93d0d8b3-f03a-4209-9ebf-1d19e41d883f service nova] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Detach interface failed, port_id=e03a2c30-fa4a-4149-9e55-2631031aa8b0, reason: Instance 67f8274b-a0ad-419a-81fc-515b06ad41aa could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1025.469874] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1025.469874] env[62000]: value = "task-882834" [ 1025.469874] env[62000]: _type = "Task" [ 1025.469874] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.478153] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.783535] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1025.814464] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882830, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555624} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.814745] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1025.814968] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1025.815262] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-857e6759-0cdb-48ce-b15d-9e7595cc377e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.823635] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1025.823635] env[62000]: value = "task-882835" [ 1025.823635] env[62000]: _type = "Task" [ 1025.823635] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.831539] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.893536] env[62000]: INFO nova.compute.manager [-] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Took 1.29 seconds to deallocate network for instance. [ 1025.910579] env[62000]: DEBUG oslo_vmware.api [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882833, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.208877} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.910843] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1025.911068] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1025.911256] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1025.911449] env[62000]: INFO nova.compute.manager [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Took 2.35 seconds to destroy the instance on the hypervisor. [ 1025.911689] env[62000]: DEBUG oslo.service.loopingcall [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.911884] env[62000]: DEBUG nova.compute.manager [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1025.911981] env[62000]: DEBUG nova.network.neutron [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1025.979811] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882834, 'name': ReconfigVM_Task, 'duration_secs': 0.140994} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.980152] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201633', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'name': 'volume-d6f79cc7-55cc-4d38-9301-08c8d382294f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'serial': 'd6f79cc7-55cc-4d38-9301-08c8d382294f'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1025.980700] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-70f8a857-a61d-4937-bb37-36f9b570075c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.987177] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1025.987177] env[62000]: value = "task-882836" [ 1025.987177] env[62000]: _type = "Task" [ 1025.987177] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.994820] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882836, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.304933] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.305213] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.306687] env[62000]: INFO nova.compute.claims [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.334020] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068294} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.334336] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1026.335103] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26116abb-c938-4fa7-b80c-a3aa0713b6ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.358541] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1026.359487] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7c7db212-0724-497c-a2ad-18d798fbbbb3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.380847] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1026.380847] env[62000]: value = "task-882837" [ 1026.380847] env[62000]: _type = "Task" [ 1026.380847] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.391374] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882837, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.400226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.498080] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882836, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.886608] env[62000]: DEBUG nova.network.neutron [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.892911] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882837, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.928140] env[62000]: DEBUG nova.compute.manager [req-71d34755-fd9f-4835-8618-c1468513c771 req-c9b2de3a-492f-4754-aff4-d8396019fb76 service nova] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Received event network-vif-deleted-1eb9c5c5-8fe1-4652-986f-3b9abdd36485 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1026.998610] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882836, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.388695] env[62000]: INFO nova.compute.manager [-] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Took 1.48 seconds to deallocate network for instance. [ 1027.394771] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882837, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.425419] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55b3d999-d495-481a-ab55-f7c549aca991 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.433319] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c60cc6b-c9c8-4f93-bbda-7ee07d9745e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.464949] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042f41e6-5a9b-454c-99c8-aea10276d53d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.473139] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c63ca3-303b-4380-9328-2e6ca2d0959f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.486919] env[62000]: DEBUG nova.compute.provider_tree [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.497187] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882836, 'name': Rename_Task, 'duration_secs': 1.134859} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.497997] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1027.498247] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7fe26504-2fee-49fa-8ed9-1355ae93d156 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.504790] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1027.504790] env[62000]: value = "task-882838" [ 1027.504790] env[62000]: _type = "Task" [ 1027.504790] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.512199] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882838, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.892827] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882837, 'name': ReconfigVM_Task, 'duration_secs': 1.284648} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.893033] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1027.893670] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a67cf2c3-abac-4d7c-8b33-f30b0ed304ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.900826] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1027.900826] env[62000]: value = "task-882839" [ 1027.900826] env[62000]: _type = "Task" [ 1027.900826] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.909092] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882839, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.941458] env[62000]: INFO nova.compute.manager [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Took 0.55 seconds to detach 1 volumes for instance. [ 1027.990338] env[62000]: DEBUG nova.scheduler.client.report [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.017064] env[62000]: DEBUG oslo_vmware.api [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882838, 'name': PowerOnVM_Task, 'duration_secs': 0.468621} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.017296] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1028.017503] env[62000]: INFO nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 9.40 seconds to spawn the instance on the hypervisor. [ 1028.017681] env[62000]: DEBUG nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.018475] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28919194-3909-45af-b8f1-6a91f1e9583d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.412042] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882839, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.448274] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.495284] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.495941] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1028.498876] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.099s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.499110] env[62000]: DEBUG nova.objects.instance [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'resources' on Instance uuid 67f8274b-a0ad-419a-81fc-515b06ad41aa {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1028.536711] env[62000]: INFO nova.compute.manager [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 17.64 seconds to build instance. [ 1028.912632] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882839, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.002532] env[62000]: DEBUG nova.compute.utils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.007051] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.007239] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1029.041238] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d692f36b-5f42-45c4-a8e9-033b6ea4032c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.151s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1029.071400] env[62000]: DEBUG nova.policy [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98729052932148a5a6b6e6c9581353f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '04d07461c67e4868a33a345d2e08db82', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.159949] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce96a38f-03b4-499d-b739-fc18ce241440 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.171923] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5f34fa3-54cf-48c0-89f4-53c78388f159 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.209883] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1eb56ec-245c-4e92-8b4e-acf7523540cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.213558] env[62000]: DEBUG nova.compute.manager [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1029.213752] env[62000]: DEBUG nova.compute.manager [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing instance network info cache due to event network-changed-adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1029.213969] env[62000]: DEBUG oslo_concurrency.lockutils [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] Acquiring lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.214132] env[62000]: DEBUG oslo_concurrency.lockutils [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] Acquired lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.214305] env[62000]: DEBUG nova.network.neutron [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Refreshing network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1029.222150] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-753a1687-fff5-4787-ae64-b17f3ca186a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.236974] env[62000]: DEBUG nova.compute.provider_tree [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1029.345038] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Successfully created port: 2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.414162] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882839, 'name': Rename_Task, 'duration_secs': 1.154088} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.414162] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1029.415267] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bf1b5aea-54ce-4ab9-ab22-c9b730c4d6ca {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.423022] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1029.423022] env[62000]: value = "task-882840" [ 1029.423022] env[62000]: _type = "Task" [ 1029.423022] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.436235] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882840, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.507544] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.740099] env[62000]: DEBUG nova.scheduler.client.report [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1029.933895] env[62000]: DEBUG oslo_vmware.api [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882840, 'name': PowerOnVM_Task, 'duration_secs': 0.489144} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.934214] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1029.934465] env[62000]: INFO nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Took 9.05 seconds to spawn the instance on the hypervisor. [ 1029.934655] env[62000]: DEBUG nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1029.935609] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5167b7be-35e6-4be8-bbb4-ce98a03ca804 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.965353] env[62000]: DEBUG nova.network.neutron [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updated VIF entry in instance network info cache for port adbea477-8af6-4cb5-82e6-1292553013f2. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1029.965353] env[62000]: DEBUG nova.network.neutron [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [{"id": "adbea477-8af6-4cb5-82e6-1292553013f2", "address": "fa:16:3e:98:64:ef", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapadbea477-8a", "ovs_interfaceid": "adbea477-8af6-4cb5-82e6-1292553013f2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.245205] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.746s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.247558] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.799s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.247809] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.265982] env[62000]: INFO nova.scheduler.client.report [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted allocations for instance 3878579f-6435-4fe3-9f8c-8461d8ac57ee [ 1030.267763] env[62000]: INFO nova.scheduler.client.report [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance 67f8274b-a0ad-419a-81fc-515b06ad41aa [ 1030.457555] env[62000]: INFO nova.compute.manager [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Took 18.90 seconds to build instance. [ 1030.469577] env[62000]: DEBUG oslo_concurrency.lockutils [req-48c19149-60fe-4907-a96d-b7bc7d3c57a0 req-7c96d3c6-c5b9-46a2-bd51-a983bcedf03c service nova] Releasing lock "refresh_cache-f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.520329] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.547218] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.547570] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.547755] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.547945] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.548117] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.548277] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.548502] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.548692] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.548908] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.549091] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.549275] env[62000]: DEBUG nova.virt.hardware [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.550206] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d92a5d3-60b1-4b00-8615-b6bc8865d047 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.559066] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db65cf9e-93e1-459f-8d19-064c957fe10b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.670409] env[62000]: DEBUG nova.compute.manager [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1030.778386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-49f2830b-e7a9-4c1e-8e6b-5e9462e30ef8 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "3878579f-6435-4fe3-9f8c-8461d8ac57ee" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.241s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.779571] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9b0d2d0c-001b-49e6-8fbc-17ba753f2bee tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "67f8274b-a0ad-419a-81fc-515b06ad41aa" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.786s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.959965] env[62000]: DEBUG oslo_concurrency.lockutils [None req-e7b4f7a1-a80a-4949-9ab6-055b9c2789ed tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.415s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.194512] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.194854] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.278629] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Successfully updated port: 2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.285805] env[62000]: DEBUG nova.compute.manager [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Received event network-vif-plugged-2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.286238] env[62000]: DEBUG oslo_concurrency.lockutils [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] Acquiring lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.286479] env[62000]: DEBUG oslo_concurrency.lockutils [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.286479] env[62000]: DEBUG oslo_concurrency.lockutils [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.286628] env[62000]: DEBUG nova.compute.manager [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] No waiting events found dispatching network-vif-plugged-2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.287284] env[62000]: WARNING nova.compute.manager [req-0a7dc324-3ef9-41f0-8594-d97d25e79f2e req-b480b5f2-a0c6-48ae-b7f1-3d99e27ad93b service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Received unexpected event network-vif-plugged-2eeb021a-8f1b-4da4-833b-fbb659424739 for instance with vm_state building and task_state spawning. [ 1031.329099] env[62000]: DEBUG nova.compute.manager [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Received event network-changed-9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1031.329337] env[62000]: DEBUG nova.compute.manager [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Refreshing instance network info cache due to event network-changed-9f791142-c29f-4dda-b815-14d5e18a8f8c. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1031.329482] env[62000]: DEBUG oslo_concurrency.lockutils [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] Acquiring lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.329628] env[62000]: DEBUG oslo_concurrency.lockutils [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] Acquired lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.329794] env[62000]: DEBUG nova.network.neutron [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Refreshing network info cache for port 9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1031.700713] env[62000]: INFO nova.compute.claims [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1031.781443] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.781601] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.781758] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1031.951382] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.951615] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.064804] env[62000]: DEBUG nova.network.neutron [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updated VIF entry in instance network info cache for port 9f791142-c29f-4dda-b815-14d5e18a8f8c. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1032.065221] env[62000]: DEBUG nova.network.neutron [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [{"id": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "address": "fa:16:3e:ee:fd:8d", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f791142-c2", "ovs_interfaceid": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.207349] env[62000]: INFO nova.compute.resource_tracker [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating resource usage from migration 4381a5f3-f842-4061-8e43-54f328315ae0 [ 1032.246662] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.246933] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.314659] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1032.354351] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af2afff3-d50b-474d-af1c-827b7f576553 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.365055] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb45ac4c-316b-471c-9b89-357f99fbab6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.398845] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9623e2e-beeb-4baf-8ba0-cc15b659444a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.407278] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d97bc54-7eef-4674-9e42-ecbea984270f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.421242] env[62000]: DEBUG nova.compute.provider_tree [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.454291] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.481225] env[62000]: DEBUG nova.network.neutron [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Updating instance_info_cache with network_info: [{"id": "2eeb021a-8f1b-4da4-833b-fbb659424739", "address": "fa:16:3e:a4:16:a5", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eeb021a-8f", "ovs_interfaceid": "2eeb021a-8f1b-4da4-833b-fbb659424739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.567962] env[62000]: DEBUG oslo_concurrency.lockutils [req-b3650f39-1cd2-40b7-81ca-55dede7046ca req-c67a9909-9a86-4ffc-abf5-794cdabc7d04 service nova] Releasing lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.749665] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.927047] env[62000]: DEBUG nova.scheduler.client.report [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.971617] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.983637] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.983950] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Instance network_info: |[{"id": "2eeb021a-8f1b-4da4-833b-fbb659424739", "address": "fa:16:3e:a4:16:a5", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eeb021a-8f", "ovs_interfaceid": "2eeb021a-8f1b-4da4-833b-fbb659424739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1032.984437] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a4:16:a5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '25f42474-5594-4733-a681-6c69f4afb946', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2eeb021a-8f1b-4da4-833b-fbb659424739', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1032.991722] env[62000]: DEBUG oslo.service.loopingcall [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1032.991939] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1032.992205] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2d18ce7e-1d3e-437e-877f-4fe82bee9ddf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.014130] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.014130] env[62000]: value = "task-882841" [ 1033.014130] env[62000]: _type = "Task" [ 1033.014130] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.024925] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882841, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.271905] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.315968] env[62000]: DEBUG nova.compute.manager [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Received event network-changed-2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.316154] env[62000]: DEBUG nova.compute.manager [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Refreshing instance network info cache due to event network-changed-2eeb021a-8f1b-4da4-833b-fbb659424739. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.316382] env[62000]: DEBUG oslo_concurrency.lockutils [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] Acquiring lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.316558] env[62000]: DEBUG oslo_concurrency.lockutils [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] Acquired lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.316731] env[62000]: DEBUG nova.network.neutron [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Refreshing network info cache for port 2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.359925] env[62000]: DEBUG nova.compute.manager [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1033.360235] env[62000]: DEBUG nova.compute.manager [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing instance network info cache due to event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1033.360646] env[62000]: DEBUG oslo_concurrency.lockutils [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.360685] env[62000]: DEBUG oslo_concurrency.lockutils [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.360843] env[62000]: DEBUG nova.network.neutron [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1033.432518] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.237s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.432879] env[62000]: INFO nova.compute.manager [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Migrating [ 1033.444528] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.473s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.447113] env[62000]: INFO nova.compute.claims [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1033.525649] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882841, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.962094] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.962392] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquired lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.962499] env[62000]: DEBUG nova.network.neutron [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1034.028540] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882841, 'name': CreateVM_Task, 'duration_secs': 0.615006} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.028718] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1034.029382] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.029554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.029881] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.030257] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-193a9d63-d8f7-402a-abb4-0349c171d88f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.035329] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1034.035329] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977d76-f336-0552-39d6-d5b68453fdfa" [ 1034.035329] env[62000]: _type = "Task" [ 1034.035329] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.043819] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977d76-f336-0552-39d6-d5b68453fdfa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.145322] env[62000]: DEBUG nova.network.neutron [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Updated VIF entry in instance network info cache for port 2eeb021a-8f1b-4da4-833b-fbb659424739. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.145708] env[62000]: DEBUG nova.network.neutron [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Updating instance_info_cache with network_info: [{"id": "2eeb021a-8f1b-4da4-833b-fbb659424739", "address": "fa:16:3e:a4:16:a5", "network": {"id": "ac4e9c3e-a4ae-475a-b5ed-69937267ab8e", "bridge": "br-int", "label": "tempest-ServersTestJSON-1534649712-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "04d07461c67e4868a33a345d2e08db82", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "25f42474-5594-4733-a681-6c69f4afb946", "external-id": "nsx-vlan-transportzone-453", "segmentation_id": 453, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2eeb021a-8f", "ovs_interfaceid": "2eeb021a-8f1b-4da4-833b-fbb659424739", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.184898] env[62000]: DEBUG nova.network.neutron [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updated VIF entry in instance network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1034.185341] env[62000]: DEBUG nova.network.neutron [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.549460] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52977d76-f336-0552-39d6-d5b68453fdfa, 'name': SearchDatastore_Task, 'duration_secs': 0.031717} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.549797] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.550103] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.550360] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.550508] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.550693] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.550985] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4b0fdef0-8fc6-40e3-87d7-81f37d390a8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.566656] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.566865] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1034.569745] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7439026-1ad2-4a17-bd98-52a540b4c4fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.575548] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1034.575548] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6eefa-aae3-0c68-508c-ea713469b847" [ 1034.575548] env[62000]: _type = "Task" [ 1034.575548] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.583676] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6eefa-aae3-0c68-508c-ea713469b847, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.619243] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eda8bc4d-24fd-4d4a-83bc-f9340b69181c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.627259] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f466c0d-2f07-4480-a953-111f36aff12e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.656404] env[62000]: DEBUG oslo_concurrency.lockutils [req-058a65f2-0f2f-4cb6-80aa-315a6f3836fb req-1d72c819-e979-4aea-80e2-f3a768bb9eb9 service nova] Releasing lock "refresh_cache-67af85f1-9134-440d-a5df-09ec7d3e72a0" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.659581] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23a7851-5e0c-48dc-b6fb-b4a1dfd2fa24 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.667096] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70850b5-c250-44b8-bd74-9819c90bd190 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.680499] env[62000]: DEBUG nova.compute.provider_tree [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1034.687373] env[62000]: DEBUG oslo_concurrency.lockutils [req-a31436e7-f302-4beb-8832-91bc816b0a08 req-91e397c1-8e31-4772-9e14-10ea62d8c482 service nova] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.724900] env[62000]: DEBUG nova.network.neutron [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [{"id": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "address": "fa:16:3e:ee:fd:8d", "network": {"id": "1e966db4-f253-40ed-b80c-38e2fe2956be", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherA-902453978-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.139", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6106a70abdad4111977eebbd82434337", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e044cfd4-1b0d-4d88-b1bd-604025731d3f", "external-id": "nsx-vlan-transportzone-372", "segmentation_id": 372, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9f791142-c2", "ovs_interfaceid": "9f791142-c29f-4dda-b815-14d5e18a8f8c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.086398] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d6eefa-aae3-0c68-508c-ea713469b847, 'name': SearchDatastore_Task, 'duration_secs': 0.014235} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.086857] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12bee46b-d012-4242-89a2-796bbb1bf203 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.092245] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1035.092245] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee4982-3753-c7f1-ac50-58018432a54a" [ 1035.092245] env[62000]: _type = "Task" [ 1035.092245] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.100066] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee4982-3753-c7f1-ac50-58018432a54a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.183456] env[62000]: DEBUG nova.scheduler.client.report [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1035.228022] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Releasing lock "refresh_cache-9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.574752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.575251] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.603754] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ee4982-3753-c7f1-ac50-58018432a54a, 'name': SearchDatastore_Task, 'duration_secs': 0.010113} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.603754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.603754] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 67af85f1-9134-440d-a5df-09ec7d3e72a0/67af85f1-9134-440d-a5df-09ec7d3e72a0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1035.604014] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e8f0280f-b57a-462a-a1da-fd835fc0e057 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.612339] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1035.612339] env[62000]: value = "task-882842" [ 1035.612339] env[62000]: _type = "Task" [ 1035.612339] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.620740] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882842, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.688926] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.244s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.689607] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1035.692953] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.421s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.694682] env[62000]: INFO nova.compute.claims [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1036.078871] env[62000]: DEBUG nova.compute.utils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.122592] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882842, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500109} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.123061] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 67af85f1-9134-440d-a5df-09ec7d3e72a0/67af85f1-9134-440d-a5df-09ec7d3e72a0.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1036.123291] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1036.123833] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e7f32cb3-ac11-4cf3-a7bc-7e54d88a84fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.130576] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1036.130576] env[62000]: value = "task-882843" [ 1036.130576] env[62000]: _type = "Task" [ 1036.130576] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.137815] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882843, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.200480] env[62000]: DEBUG nova.compute.utils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1036.204583] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1036.204751] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1036.244370] env[62000]: DEBUG nova.policy [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c3018dc2fa8d413aa464bf4e0c32341e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d131f95ad3949d89cd6f36f6648d3f1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1036.554434] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Successfully created port: 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1036.582291] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.640339] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882843, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064723} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.640623] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1036.641418] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf06421-5e42-4578-9866-ad629b92794c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.665894] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] 67af85f1-9134-440d-a5df-09ec7d3e72a0/67af85f1-9134-440d-a5df-09ec7d3e72a0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1036.666224] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3593203d-ec81-44dc-a142-9fbffe935092 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.685328] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1036.685328] env[62000]: value = "task-882844" [ 1036.685328] env[62000]: _type = "Task" [ 1036.685328] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.692860] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882844, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.703786] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1036.742347] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27453f3-7192-413d-b3c4-ad934039feed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.763791] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1036.891628] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d18348-1a40-467b-86d8-fa8ca359ac77 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.899442] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1097987-3881-4ac6-b54a-2f93a5280a4f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.929281] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2896fbb-5b98-4acc-a61e-039df7f8b5fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.936657] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ae3e8f7-9782-4a10-b39f-24a2f9ce89b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.951550] env[62000]: DEBUG nova.compute.provider_tree [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.849623] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1037.849623] env[62000]: DEBUG nova.scheduler.client.report [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.854423] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1f4ee65-da98-436f-bcaa-9515cee5776b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.863773] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882844, 'name': ReconfigVM_Task, 'duration_secs': 0.716281} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.864057] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1037.864057] env[62000]: value = "task-882845" [ 1037.864057] env[62000]: _type = "Task" [ 1037.864057] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.864298] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Reconfigured VM instance instance-00000067 to attach disk [datastore1] 67af85f1-9134-440d-a5df-09ec7d3e72a0/67af85f1-9134-440d-a5df-09ec7d3e72a0.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1037.864952] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-23466e26-f475-4aef-b2d7-9afb89abbc15 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.875451] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882845, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.876887] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1037.876887] env[62000]: value = "task-882846" [ 1037.876887] env[62000]: _type = "Task" [ 1037.876887] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.885422] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882846, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.063721] env[62000]: DEBUG nova.compute.manager [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1038.063944] env[62000]: DEBUG oslo_concurrency.lockutils [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.064182] env[62000]: DEBUG oslo_concurrency.lockutils [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.064390] env[62000]: DEBUG oslo_concurrency.lockutils [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.064596] env[62000]: DEBUG nova.compute.manager [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] No waiting events found dispatching network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.064772] env[62000]: WARNING nova.compute.manager [req-3cc790ac-ecae-4e2f-9f2d-7ad1ab5f2a58 req-3add8194-ceea-439f-a35f-7b8154a68f31 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received unexpected event network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e for instance with vm_state building and task_state spawning. [ 1038.147490] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Successfully updated port: 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1038.351445] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.351713] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.351958] env[62000]: INFO nova.compute.manager [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Attaching volume 69680b18-0928-4a4e-a64e-2b8bb610c909 to /dev/sdb [ 1038.354587] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.357284] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.664s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.357768] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1038.374854] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882845, 'name': PowerOffVM_Task, 'duration_secs': 0.167023} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.376732] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1038.376934] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1038.386012] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.386254] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.386417] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.386606] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.386758] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.386909] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.387130] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.387299] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.387469] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.387635] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.387808] env[62000]: DEBUG nova.virt.hardware [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.388574] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f76b09a8-7e38-4a7b-9b07-8b970acfb595 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.395126] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882846, 'name': Rename_Task, 'duration_secs': 0.152119} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.395768] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cdda4f-f281-4f0b-86c3-9a3728adaa07 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.398358] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1038.398819] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e38b08b1-4f17-478d-99f1-185a805b6352 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.404967] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80fe4068-efc8-4900-bb1e-27315985b714 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.409858] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-833e28e3-4d5b-48fe-ab7c-e577005535ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.412352] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1038.412352] env[62000]: value = "task-882847" [ 1038.412352] env[62000]: _type = "Task" [ 1038.412352] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.427963] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882847, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.438961] env[62000]: DEBUG nova.virt.block_device [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updating existing volume attachment record: 0442003d-ea53-4257-82e5-b8bf1af6a294 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1038.650480] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1038.650638] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1038.650834] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1038.864991] env[62000]: DEBUG nova.compute.utils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.866509] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1038.866676] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1038.883445] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.883766] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.883944] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.884153] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.884361] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.884567] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.884820] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.885021] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.885206] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.885414] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.885630] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.891202] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-501554ee-24b3-41aa-98ee-60ddfd18e5c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.905241] env[62000]: DEBUG nova.policy [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7f5c46ad438d4b1eaa86c4647f20a1a7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '43e5c47870584d05abaf9de72d45cce2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.908266] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1038.908266] env[62000]: value = "task-882849" [ 1038.908266] env[62000]: _type = "Task" [ 1038.908266] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.919474] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882849, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.924136] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882847, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.208729] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1039.257194] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Successfully created port: 45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1039.369535] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1039.423877] env[62000]: DEBUG oslo_vmware.api [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882849, 'name': ReconfigVM_Task, 'duration_secs': 0.158716} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.424542] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1039.430908] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882847, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.495663] env[62000]: DEBUG nova.network.neutron [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.928128] env[62000]: DEBUG oslo_vmware.api [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882847, 'name': PowerOnVM_Task, 'duration_secs': 1.058137} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.928600] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1039.928954] env[62000]: INFO nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Took 9.41 seconds to spawn the instance on the hypervisor. [ 1039.929305] env[62000]: DEBUG nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1039.930509] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-295f70bb-58ed-41b6-ac09-110f208ff6d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.937152] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1039.937499] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1039.937762] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1039.938099] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1039.938401] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1039.938696] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1039.939069] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1039.939411] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1039.939787] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1039.940046] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1039.940380] env[62000]: DEBUG nova.virt.hardware [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.998188] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.998544] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance network_info: |[{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1039.998977] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:cc:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69c33d12-7f23-4311-89f4-5b66d2fc837e', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1040.006403] env[62000]: DEBUG oslo.service.loopingcall [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1040.006872] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1040.007089] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-47c3259f-5c77-4c36-9150-fea99e9bfce7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.027610] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1040.027610] env[62000]: value = "task-882850" [ 1040.027610] env[62000]: _type = "Task" [ 1040.027610] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.036377] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882850, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.092566] env[62000]: DEBUG nova.compute.manager [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1040.092817] env[62000]: DEBUG nova.compute.manager [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing instance network info cache due to event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1040.093073] env[62000]: DEBUG oslo_concurrency.lockutils [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.093234] env[62000]: DEBUG oslo_concurrency.lockutils [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.093421] env[62000]: DEBUG nova.network.neutron [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1040.378583] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1040.404064] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1040.404373] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1040.404564] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1040.404769] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1040.404923] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1040.405089] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1040.405305] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1040.405468] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1040.405636] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1040.405809] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1040.405978] env[62000]: DEBUG nova.virt.hardware [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1040.406848] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c558d47-1cf2-4439-a1f7-938fd6843ceb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.414501] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0443ed95-ab75-4bb8-abfd-d728f2bf586e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.453860] env[62000]: ERROR nova.compute.manager [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Setting instance vm_state to ERROR: AttributeError: 'NoneType' object has no attribute 'key' [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Traceback (most recent call last): [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/compute/manager.py", line 10863, in _error_out_instance_on_exception [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] yield [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] disk_info = self.driver.migrate_disk_and_power_off( [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] return self._vmops.migrate_disk_and_power_off(context, instance, [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] disk_key = device.key [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] AttributeError: 'NoneType' object has no attribute 'key' [ 1040.453860] env[62000]: ERROR nova.compute.manager [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] [ 1040.467927] env[62000]: INFO nova.compute.manager [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Took 14.18 seconds to build instance. [ 1040.537040] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882850, 'name': CreateVM_Task, 'duration_secs': 0.294879} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.537239] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1040.537951] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.538149] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.538475] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1040.538729] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d8e828d-ee18-45ff-8307-3c722a154cee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.543143] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1040.543143] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52446686-f0a8-520f-ae03-0db6382a9cb0" [ 1040.543143] env[62000]: _type = "Task" [ 1040.543143] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.552480] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52446686-f0a8-520f-ae03-0db6382a9cb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.937529] env[62000]: DEBUG nova.network.neutron [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updated VIF entry in instance network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1040.937888] env[62000]: DEBUG nova.network.neutron [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.969639] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a7334ca5-3acc-4252-a1f9-6c65973a8e7f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.688s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.975828] env[62000]: INFO nova.compute.manager [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Swapping old allocation on dict_keys(['bcfccf20-49dd-4b91-819e-4373e67bf5ec']) held by migration 4381a5f3-f842-4061-8e43-54f328315ae0 for instance [ 1041.004444] env[62000]: DEBUG nova.scheduler.client.report [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Overwriting current allocation {'allocations': {'bcfccf20-49dd-4b91-819e-4373e67bf5ec': {'resources': {'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 135}}, 'project_id': '6106a70abdad4111977eebbd82434337', 'user_id': '79730e91610c4c598019b122b4f41418', 'consumer_generation': 1} on consumer 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 {{(pid=62000) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1041.057547] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52446686-f0a8-520f-ae03-0db6382a9cb0, 'name': SearchDatastore_Task, 'duration_secs': 0.009961} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.057547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.057547] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1041.057547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.057547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.057547] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1041.057547] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5a5e37a0-42fd-4fe3-9970-161562da380c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.066309] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1041.066714] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1041.071266] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26e52804-7b33-4e0d-b5eb-8358c8aa961f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.079897] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1041.079897] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcb5e4-5762-2a01-7a8f-cd056181c744" [ 1041.079897] env[62000]: _type = "Task" [ 1041.079897] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.089680] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcb5e4-5762-2a01-7a8f-cd056181c744, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.117180] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Successfully updated port: 45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1041.435109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.435384] env[62000]: DEBUG oslo_concurrency.lockutils [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.435574] env[62000]: DEBUG nova.compute.manager [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1041.436562] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74a3f150-1d16-4658-bf11-769cd8a15618 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.439672] env[62000]: DEBUG oslo_concurrency.lockutils [req-8d68cdd2-0753-4212-928e-8b69e33287de req-3a58ed4b-2897-491d-9d4a-75cd7be391bd service nova] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.444160] env[62000]: DEBUG nova.compute.manager [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62000) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1041.444736] env[62000]: DEBUG nova.objects.instance [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'flavor' on Instance uuid 67af85f1-9134-440d-a5df-09ec7d3e72a0 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.590547] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52bcb5e4-5762-2a01-7a8f-cd056181c744, 'name': SearchDatastore_Task, 'duration_secs': 0.008294} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.591994] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b157bc5-fdb9-43f0-9d51-e44f7b9a68bf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.598035] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1041.598035] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5206a82a-8b0c-5436-a831-e52bfbcae49e" [ 1041.598035] env[62000]: _type = "Task" [ 1041.598035] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.604676] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5206a82a-8b0c-5436-a831-e52bfbcae49e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.619501] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1041.619646] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.619788] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1041.697357] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.697589] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.697805] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.698010] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.698204] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.700598] env[62000]: INFO nova.compute.manager [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Terminating instance [ 1041.702426] env[62000]: DEBUG nova.compute.manager [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.702642] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1041.702872] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0145607b-f95f-4074-8f4f-016249f50114 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.710428] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1041.710428] env[62000]: value = "task-882852" [ 1041.710428] env[62000]: _type = "Task" [ 1041.710428] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.718142] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882852, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.950273] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1041.950554] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e917f0f1-9f5a-4fc2-badd-9799c441c078 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.958013] env[62000]: DEBUG oslo_vmware.api [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1041.958013] env[62000]: value = "task-882853" [ 1041.958013] env[62000]: _type = "Task" [ 1041.958013] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.966387] env[62000]: DEBUG oslo_vmware.api [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882853, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.108037] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5206a82a-8b0c-5436-a831-e52bfbcae49e, 'name': SearchDatastore_Task, 'duration_secs': 0.008959} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.108314] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.108583] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1042.108843] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-51286070-9b80-4cd1-a579-e60c1efc1f18 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.115953] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1042.115953] env[62000]: value = "task-882854" [ 1042.115953] env[62000]: _type = "Task" [ 1042.115953] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.118780] env[62000]: DEBUG nova.compute.manager [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Received event network-vif-plugged-45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.118780] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.118899] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.119082] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.119142] env[62000]: DEBUG nova.compute.manager [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] No waiting events found dispatching network-vif-plugged-45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1042.119343] env[62000]: WARNING nova.compute.manager [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Received unexpected event network-vif-plugged-45862b3c-e63d-4af1-9192-270c4f66cabb for instance with vm_state building and task_state spawning. [ 1042.119532] env[62000]: DEBUG nova.compute.manager [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Received event network-changed-45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1042.119715] env[62000]: DEBUG nova.compute.manager [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Refreshing instance network info cache due to event network-changed-45862b3c-e63d-4af1-9192-270c4f66cabb. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1042.119971] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Acquiring lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.133598] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882854, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.154440] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1042.221522] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1042.221775] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1042.221980] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201633', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'name': 'volume-d6f79cc7-55cc-4d38-9301-08c8d382294f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'serial': 'd6f79cc7-55cc-4d38-9301-08c8d382294f'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1042.225078] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fab0a06c-f26e-4cd7-b1b2-9f1895ed0146 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.243306] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ac88ca3-980e-466b-97fc-2c76a8fa5a7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.249726] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9de1abf-5f29-43a3-a021-241c5b92a876 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.269614] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f17608-5dc0-45a1-b8b0-23888b67098c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.284849] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] The volume has not been displaced from its original location: [datastore1] volume-d6f79cc7-55cc-4d38-9301-08c8d382294f/volume-d6f79cc7-55cc-4d38-9301-08c8d382294f.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1042.290271] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1042.290640] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-351539c5-8e75-4731-87ad-1b2113055189 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.306602] env[62000]: DEBUG nova.network.neutron [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.309308] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1042.309308] env[62000]: value = "task-882855" [ 1042.309308] env[62000]: _type = "Task" [ 1042.309308] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.317716] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882855, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.468169] env[62000]: DEBUG oslo_vmware.api [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882853, 'name': PowerOffVM_Task, 'duration_secs': 0.17968} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.468477] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1042.468663] env[62000]: DEBUG nova.compute.manager [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.469479] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90df29b-32d4-44a0-836c-0fd0b04aa168 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.585730] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.586052] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.629986] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882854, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.446204} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.630279] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1042.630495] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1042.630746] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-25899da4-c8df-4bf6-8d83-b7f601402ce2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.636475] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1042.636475] env[62000]: value = "task-882856" [ 1042.636475] env[62000]: _type = "Task" [ 1042.636475] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.644065] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882856, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.722650] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fc5f92-bf2a-4563-9889-12f1ec39778b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.729858] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-630ef509-1760-4aac-aae3-40d4b414c381 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.758391] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29930b39-9109-4110-bc07-ee28615025b0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.765216] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d5c85f1-23ad-443e-8b80-7aab41a9753b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.777861] env[62000]: DEBUG nova.compute.provider_tree [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.810522] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.810808] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Instance network_info: |[{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.811186] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Acquired lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.811392] env[62000]: DEBUG nova.network.neutron [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Refreshing network info cache for port 45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1042.812877] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fb:b2:f5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '68ec9c06-8680-4a41-abad-cddbd1f768c9', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '45862b3c-e63d-4af1-9192-270c4f66cabb', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.819566] env[62000]: DEBUG oslo.service.loopingcall [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.823273] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1042.823499] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d854392-2d4a-40c5-9c9d-d839edddbcd6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.842968] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882855, 'name': ReconfigVM_Task, 'duration_secs': 0.337347} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.844090] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1042.848368] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.848368] env[62000]: value = "task-882857" [ 1042.848368] env[62000]: _type = "Task" [ 1042.848368] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.848755] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8448c040-4bf1-4aa7-85f7-fb23d1e08699 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.866930] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882857, 'name': CreateVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.868071] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1042.868071] env[62000]: value = "task-882858" [ 1042.868071] env[62000]: _type = "Task" [ 1042.868071] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.875602] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882858, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.982969] env[62000]: DEBUG oslo_concurrency.lockutils [None req-20e2c932-6551-40c0-ab87-48e0d42f2c8f tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.985062] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1042.985182] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201646', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'name': 'volume-69680b18-0928-4a4e-a64e-2b8bb610c909', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0246b032-505e-4bc5-bfc0-5779ff564626', 'attached_at': '', 'detached_at': '', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'serial': '69680b18-0928-4a4e-a64e-2b8bb610c909'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1042.986107] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c388d8da-5624-4d8c-8210-ccc1cebc0f75 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.004497] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b998586-443b-4299-9c55-a8f2daa6bb0b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.021833] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.029111] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] volume-69680b18-0928-4a4e-a64e-2b8bb610c909/volume-69680b18-0928-4a4e-a64e-2b8bb610c909.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.029636] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e11b0500-a83e-40f3-a874-68b812d4fb71 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.047549] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1043.047549] env[62000]: value = "task-882859" [ 1043.047549] env[62000]: _type = "Task" [ 1043.047549] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.055089] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.145881] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882856, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061004} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.146249] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1043.146994] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2661e637-afcc-4678-8f54-ec6b16fed77e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.171487] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1043.171790] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2001ea9f-b52d-448d-9815-1c50f5e4096a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.190832] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1043.190832] env[62000]: value = "task-882860" [ 1043.190832] env[62000]: _type = "Task" [ 1043.190832] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.198129] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.281468] env[62000]: DEBUG nova.scheduler.client.report [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.369840] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882857, 'name': CreateVM_Task, 'duration_secs': 0.328954} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.373242] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1043.374089] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.374276] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.374642] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.375371] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5855aa0-830f-4e1d-ac76-a3157c3c22b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.381232] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882858, 'name': ReconfigVM_Task, 'duration_secs': 0.133574} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.381232] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201633', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'name': 'volume-d6f79cc7-55cc-4d38-9301-08c8d382294f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ee49e0b-6d37-4826-8d8a-bfb3752af4f8', 'attached_at': '', 'detached_at': '', 'volume_id': 'd6f79cc7-55cc-4d38-9301-08c8d382294f', 'serial': 'd6f79cc7-55cc-4d38-9301-08c8d382294f'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1043.381456] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1043.382319] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb7f2a39-848b-4f21-81b0-db16491d0c37 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.387014] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1043.387014] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e3ab38-f7dc-a3d7-a9f1-fdbbbdf9c181" [ 1043.387014] env[62000]: _type = "Task" [ 1043.387014] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.394274] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1043.394863] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df21d720-7186-4459-aed5-dad9be84f3f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.400381] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e3ab38-f7dc-a3d7-a9f1-fdbbbdf9c181, 'name': SearchDatastore_Task, 'duration_secs': 0.009228} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.401033] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.401303] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.401589] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.401799] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.402063] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.402327] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4573f112-a355-4c16-bea3-7f5a0f89442c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.409623] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.409804] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1043.412392] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca2ed1ff-5dca-4e91-b822-1d47adcc4ad7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.417628] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1043.417628] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216afc2-6acc-b1a9-cbb7-a5ef8fa2a527" [ 1043.417628] env[62000]: _type = "Task" [ 1043.417628] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.425191] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216afc2-6acc-b1a9-cbb7-a5ef8fa2a527, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.474916] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1043.475269] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1043.475477] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1043.475754] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8de8343-b248-4bc5-9b69-da48098c589e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.482049] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1043.482049] env[62000]: value = "task-882862" [ 1043.482049] env[62000]: _type = "Task" [ 1043.482049] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.490398] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882862, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.558521] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882859, 'name': ReconfigVM_Task, 'duration_secs': 0.438086} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.559372] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfigured VM instance instance-0000005f to attach disk [datastore1] volume-69680b18-0928-4a4e-a64e-2b8bb610c909/volume-69680b18-0928-4a4e-a64e-2b8bb610c909.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.565824] env[62000]: DEBUG nova.network.neutron [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updated VIF entry in instance network info cache for port 45862b3c-e63d-4af1-9192-270c4f66cabb. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1043.566195] env[62000]: DEBUG nova.network.neutron [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.567338] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-18cef5c3-8be9-4711-9bbd-10b5fe2fe692 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.582010] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1043.582010] env[62000]: value = "task-882863" [ 1043.582010] env[62000]: _type = "Task" [ 1043.582010] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.590845] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882863, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.702431] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882860, 'name': ReconfigVM_Task, 'duration_secs': 0.34446} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.702962] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1043.703660] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76305ca9-e3ad-48ad-8942-93cb0dd9d83a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.711024] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1043.711024] env[62000]: value = "task-882864" [ 1043.711024] env[62000]: _type = "Task" [ 1043.711024] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.719602] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882864, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.786836] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.201s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.787088] env[62000]: INFO nova.compute.manager [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Successfully reverted task state from resize_migrating on failure for instance. [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server [None req-8f74ba28-a8e0-4ab0-a927-5bf74843729b tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Exception during message handling: AttributeError: 'NoneType' object has no attribute 'key' [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server raise self.value [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server raise self.value [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server raise self.value [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6071, in resize_instance [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server raise self.value [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6068, in resize_instance [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self._resize_instance(context, instance, image, migration, [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 6105, in _resize_instance [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server disk_info = self.driver.migrate_disk_and_power_off( [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 264, in migrate_disk_and_power_off [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server return self._vmops.migrate_disk_and_power_off(context, instance, [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1467, in migrate_disk_and_power_off [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self._resize_disk(instance, vm_ref, vmdk, flavor) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1398, in _resize_disk [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server self._volumeops.detach_disk_from_vm(vm_ref, instance, vmdk.device) [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 121, in detach_disk_from_vm [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server disk_key = device.key [ 1043.795926] env[62000]: ERROR oslo_messaging.rpc.server AttributeError: 'NoneType' object has no attribute 'key' [ 1043.797710] env[62000]: ERROR oslo_messaging.rpc.server [ 1043.928743] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5216afc2-6acc-b1a9-cbb7-a5ef8fa2a527, 'name': SearchDatastore_Task, 'duration_secs': 0.009644} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.929548] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-593211df-0f24-40e6-b0b7-dbdb4465228d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.935561] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1043.935561] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e3e008-3456-b32e-0c9f-e2966e267884" [ 1043.935561] env[62000]: _type = "Task" [ 1043.935561] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.944878] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e3e008-3456-b32e-0c9f-e2966e267884, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.992468] env[62000]: DEBUG oslo_vmware.api [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882862, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.098221} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.992825] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1043.992863] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1043.993060] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1043.993252] env[62000]: INFO nova.compute.manager [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 2.29 seconds to destroy the instance on the hypervisor. [ 1043.993540] env[62000]: DEBUG oslo.service.loopingcall [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1043.993686] env[62000]: DEBUG nova.compute.manager [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1043.993782] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1044.078475] env[62000]: DEBUG oslo_concurrency.lockutils [req-0ae8826f-8965-404f-81b8-ce3b2475d748 req-20f8ba9e-1087-44ee-90ac-a4a061d8cf50 service nova] Releasing lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.091714] env[62000]: DEBUG oslo_vmware.api [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882863, 'name': ReconfigVM_Task, 'duration_secs': 0.145017} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.092506] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201646', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'name': 'volume-69680b18-0928-4a4e-a64e-2b8bb610c909', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0246b032-505e-4bc5-bfc0-5779ff564626', 'attached_at': '', 'detached_at': '', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'serial': '69680b18-0928-4a4e-a64e-2b8bb610c909'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1044.203627] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.203891] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.204124] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.204332] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.204508] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.206867] env[62000]: INFO nova.compute.manager [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Terminating instance [ 1044.209051] env[62000]: DEBUG nova.compute.manager [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.209144] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.210014] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aec6017-aa63-4733-95cd-11223b8beec9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.222234] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882864, 'name': Rename_Task, 'duration_secs': 0.135256} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.224787] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1044.225164] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1044.225407] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b95d494a-b4d5-4133-8948-39127f6fb5b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.227253] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f960d5df-8d2b-4583-a7b6-5cd44cb66f8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.235045] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1044.235045] env[62000]: value = "task-882865" [ 1044.235045] env[62000]: _type = "Task" [ 1044.235045] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.248099] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882865, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.296544] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1044.296744] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1044.297421] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore1] 67af85f1-9134-440d-a5df-09ec7d3e72a0 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.297421] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a44cc3b2-fec3-4d97-b219-b926d97e9405 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.305364] env[62000]: DEBUG oslo_vmware.api [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1044.305364] env[62000]: value = "task-882867" [ 1044.305364] env[62000]: _type = "Task" [ 1044.305364] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.315379] env[62000]: DEBUG oslo_vmware.api [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882867, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.377026] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.377384] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.377646] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.377847] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1044.378032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1044.380445] env[62000]: INFO nova.compute.manager [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Terminating instance [ 1044.382581] env[62000]: DEBUG nova.compute.manager [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1044.382790] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1044.383666] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63d16ec-d8fe-428f-9111-682fd856e2d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.391625] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1044.391910] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7e243c95-838f-4afe-8f9c-146d11860185 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.398604] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1044.398604] env[62000]: value = "task-882868" [ 1044.398604] env[62000]: _type = "Task" [ 1044.398604] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.406832] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882868, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.445717] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52e3e008-3456-b32e-0c9f-e2966e267884, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.445996] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.446308] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1044.446580] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b9ad3273-9149-43aa-aa6e-a9bf3eacfd55 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.453892] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1044.453892] env[62000]: value = "task-882869" [ 1044.453892] env[62000]: _type = "Task" [ 1044.453892] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.462301] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.517224] env[62000]: DEBUG nova.compute.manager [req-85612859-ab7e-467e-991f-041275611020 req-2e202ea6-e1d6-4f22-9f59-29a137741b7d service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Received event network-vif-deleted-9f791142-c29f-4dda-b815-14d5e18a8f8c {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1044.517224] env[62000]: INFO nova.compute.manager [req-85612859-ab7e-467e-991f-041275611020 req-2e202ea6-e1d6-4f22-9f59-29a137741b7d service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Neutron deleted interface 9f791142-c29f-4dda-b815-14d5e18a8f8c; detaching it from the instance and deleting it from the info cache [ 1044.517224] env[62000]: DEBUG nova.network.neutron [req-85612859-ab7e-467e-991f-041275611020 req-2e202ea6-e1d6-4f22-9f59-29a137741b7d service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.747319] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882865, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.815682] env[62000]: DEBUG oslo_vmware.api [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882867, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.220357} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.816157] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.816389] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1044.816624] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1044.816818] env[62000]: INFO nova.compute.manager [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1044.817093] env[62000]: DEBUG oslo.service.loopingcall [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.817367] env[62000]: DEBUG nova.compute.manager [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.817496] env[62000]: DEBUG nova.network.neutron [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1044.912877] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882868, 'name': PowerOffVM_Task, 'duration_secs': 0.189524} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.913326] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1044.913533] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1044.913896] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a3a5283b-3fe0-4be4-815b-d03d70a543f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.964730] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882869, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.982450] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.021798] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f929638b-e99a-4563-afad-6abe92f4e920 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.031635] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c350439-f784-4d85-a334-83bd329934f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.062920] env[62000]: DEBUG nova.compute.manager [req-85612859-ab7e-467e-991f-041275611020 req-2e202ea6-e1d6-4f22-9f59-29a137741b7d service nova] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Detach interface failed, port_id=9f791142-c29f-4dda-b815-14d5e18a8f8c, reason: Instance 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1045.140220] env[62000]: DEBUG nova.objects.instance [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 0246b032-505e-4bc5-bfc0-5779ff564626 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.237382] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1045.237620] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1045.237895] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] f8f26be0-f7d1-43e9-886d-c19d385e6935 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1045.241535] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6c20e3b8-f4b9-4ae1-b67d-b5c14a5e6221 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.248971] env[62000]: DEBUG oslo_vmware.api [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882865, 'name': PowerOnVM_Task, 'duration_secs': 0.622551} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.250261] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1045.250482] env[62000]: INFO nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Took 6.90 seconds to spawn the instance on the hypervisor. [ 1045.250667] env[62000]: DEBUG nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1045.251319] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1045.251319] env[62000]: value = "task-882871" [ 1045.251319] env[62000]: _type = "Task" [ 1045.251319] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.251735] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbbd6e1c-34fc-4f23-8fa2-9d581bc41a90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.266888] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.464924] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.617965} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.465247] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1045.465440] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1045.465713] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06e5c788-78f0-4a44-bc49-9fa91a207995 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.473366] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1045.473366] env[62000]: value = "task-882872" [ 1045.473366] env[62000]: _type = "Task" [ 1045.473366] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.486405] env[62000]: INFO nova.compute.manager [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 1.49 seconds to deallocate network for instance. [ 1045.486751] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882872, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.646877] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f93891b5-d490-4922-992c-5513340b6197 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.295s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.671497] env[62000]: DEBUG nova.network.neutron [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.766885] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882871, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.774885] env[62000]: INFO nova.compute.manager [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Took 12.82 seconds to build instance. [ 1045.983999] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882872, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.307874} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.984325] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.985133] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-621a1671-1f06-42cc-8ed8-2abf6cfb8fce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.006592] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.006941] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2db3646b-016c-46fa-8d3a-c3b403efb29d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.026665] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1046.026665] env[62000]: value = "task-882873" [ 1046.026665] env[62000]: _type = "Task" [ 1046.026665] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.031999] env[62000]: INFO nova.compute.manager [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 0.54 seconds to detach 1 volumes for instance. [ 1046.036558] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882873, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.037016] env[62000]: DEBUG nova.compute.manager [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deleting volume: d6f79cc7-55cc-4d38-9301-08c8d382294f {{(pid=62000) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1046.173471] env[62000]: INFO nova.compute.manager [-] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Took 1.36 seconds to deallocate network for instance. [ 1046.203922] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.204861] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.267687] env[62000]: DEBUG oslo_vmware.api [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882871, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.562337} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.267687] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1046.267687] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1046.267687] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1046.267687] env[62000]: INFO nova.compute.manager [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Took 1.88 seconds to destroy the instance on the hypervisor. [ 1046.267687] env[62000]: DEBUG oslo.service.loopingcall [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1046.268021] env[62000]: DEBUG nova.compute.manager [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1046.268149] env[62000]: DEBUG nova.network.neutron [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1046.276623] env[62000]: DEBUG oslo_concurrency.lockutils [None req-06b09308-3198-4e96-b28b-81157d1c3c34 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.325s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.536429] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882873, 'name': ReconfigVM_Task, 'duration_secs': 0.344453} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.537029] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.537465] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87e6c7d9-d726-4d6c-951e-960794a09ab3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.544713] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1046.544713] env[62000]: value = "task-882875" [ 1046.544713] env[62000]: _type = "Task" [ 1046.544713] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.554355] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882875, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.560914] env[62000]: DEBUG nova.compute.manager [req-396b5bfb-724c-496e-b559-7fdb90429961 req-d6044cfb-4bcb-4a83-9d05-936f388430af service nova] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Received event network-vif-deleted-2eeb021a-8f1b-4da4-833b-fbb659424739 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1046.576419] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.576766] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.577068] env[62000]: DEBUG nova.objects.instance [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'resources' on Instance uuid 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.680853] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.709369] env[62000]: INFO nova.compute.manager [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Detaching volume 69680b18-0928-4a4e-a64e-2b8bb610c909 [ 1046.746072] env[62000]: INFO nova.virt.block_device [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Attempting to driver detach volume 69680b18-0928-4a4e-a64e-2b8bb610c909 from mountpoint /dev/sdb [ 1046.746685] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1046.746979] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201646', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'name': 'volume-69680b18-0928-4a4e-a64e-2b8bb610c909', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0246b032-505e-4bc5-bfc0-5779ff564626', 'attached_at': '', 'detached_at': '', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'serial': '69680b18-0928-4a4e-a64e-2b8bb610c909'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1046.748020] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcc59216-7daf-4cfb-bf9e-9788be60deee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.772599] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45370032-448b-4343-8077-dc3a7905e1f7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.780340] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33bd01f5-3a67-493d-be81-2e2bc266a5df {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.800211] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-280c70b1-e688-429c-a5e0-a33f6149322e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.816167] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] The volume has not been displaced from its original location: [datastore1] volume-69680b18-0928-4a4e-a64e-2b8bb610c909/volume-69680b18-0928-4a4e-a64e-2b8bb610c909.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1046.822312] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfiguring VM instance instance-0000005f to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1046.822681] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0981090c-24c3-4419-97b0-f551ddc41f7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.844979] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1046.844979] env[62000]: value = "task-882876" [ 1046.844979] env[62000]: _type = "Task" [ 1046.844979] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.853614] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882876, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.055352] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882875, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.223024] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47a37d89-3b14-48fd-a5de-78e634c61e00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.230718] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ef74e5-e84b-4a27-91da-1b7e3c33e689 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.261300] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a56738d6-215b-4149-b603-39bda3aca56f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.268074] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7591b187-ea22-493e-b8ad-3ea333231d3d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.280655] env[62000]: DEBUG nova.compute.provider_tree [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.354843] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.384488] env[62000]: DEBUG nova.network.neutron [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1047.555515] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882875, 'name': Rename_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.702503] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.703200] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.784137] env[62000]: DEBUG nova.scheduler.client.report [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1047.855669] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882876, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.887378] env[62000]: INFO nova.compute.manager [-] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Took 1.62 seconds to deallocate network for instance. [ 1048.056233] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882875, 'name': Rename_Task, 'duration_secs': 1.148923} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.056564] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1048.056778] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0f8703f5-c998-4786-913f-e46dc39a75ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.062965] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1048.062965] env[62000]: value = "task-882877" [ 1048.062965] env[62000]: _type = "Task" [ 1048.062965] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.071517] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882877, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.206756] env[62000]: DEBUG nova.compute.utils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1048.289371] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.712s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.291584] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.611s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.291826] env[62000]: DEBUG nova.objects.instance [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid 67af85f1-9134-440d-a5df-09ec7d3e72a0 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1048.313222] env[62000]: INFO nova.scheduler.client.report [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted allocations for instance 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 [ 1048.356382] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882876, 'name': ReconfigVM_Task, 'duration_secs': 1.06678} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.357202] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Reconfigured VM instance instance-0000005f to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1048.362015] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b01916e-7a80-4343-b9f2-4043b9eaae14 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.377251] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1048.377251] env[62000]: value = "task-882878" [ 1048.377251] env[62000]: _type = "Task" [ 1048.377251] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.388034] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882878, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.393347] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.573262] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882877, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.589239] env[62000]: DEBUG nova.compute.manager [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1048.589446] env[62000]: DEBUG nova.compute.manager [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing instance network info cache due to event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1048.589671] env[62000]: DEBUG oslo_concurrency.lockutils [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.589827] env[62000]: DEBUG oslo_concurrency.lockutils [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.589982] env[62000]: DEBUG nova.network.neutron [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1048.709419] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.821652] env[62000]: DEBUG oslo_concurrency.lockutils [None req-c5adae3e-9d42-4e3a-9910-7660454f887c tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.124s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.822721] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.801s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.822957] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.823184] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.823388] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.827527] env[62000]: INFO nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Terminating instance [ 1048.831219] env[62000]: DEBUG nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.831874] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb1717b5-14fc-476e-9370-a1325428f4f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.845485] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aadf87d-b48d-4630-b09f-6bce19cb69b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.878252] env[62000]: WARNING nova.virt.vmwareapi.driver [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 could not be found. [ 1048.878524] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1048.881629] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3203ba6f-0cb6-46c4-91ef-0059cb7b68e4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.891015] env[62000]: DEBUG oslo_vmware.api [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882878, 'name': ReconfigVM_Task, 'duration_secs': 0.163177} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.892390] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201646', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'name': 'volume-69680b18-0928-4a4e-a64e-2b8bb610c909', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '0246b032-505e-4bc5-bfc0-5779ff564626', 'attached_at': '', 'detached_at': '', 'volume_id': '69680b18-0928-4a4e-a64e-2b8bb610c909', 'serial': '69680b18-0928-4a4e-a64e-2b8bb610c909'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1048.897211] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb65a11b-94a9-469a-abd2-0b4ecaddf4b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.930684] env[62000]: WARNING nova.virt.vmwareapi.vmops [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8 could not be found. [ 1048.930910] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1048.931118] env[62000]: INFO nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 0.10 seconds to destroy the instance on the hypervisor. [ 1048.931373] env[62000]: DEBUG oslo.service.loopingcall [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.933864] env[62000]: DEBUG nova.compute.manager [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1048.933864] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1048.945507] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f8cbfc7-16e1-46d2-b775-531f65da0698 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.953122] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f71328a-3b6c-4a7a-a309-ad33cc2bbb39 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.983367] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17f40570-891f-441f-95b2-de0276e7038a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.990107] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d58594f9-eddf-4cff-b48a-70c038b1191a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.002888] env[62000]: DEBUG nova.compute.provider_tree [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1049.073734] env[62000]: DEBUG oslo_vmware.api [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882877, 'name': PowerOnVM_Task, 'duration_secs': 0.858584} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.074092] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1049.074159] env[62000]: INFO nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Took 8.70 seconds to spawn the instance on the hypervisor. [ 1049.074347] env[62000]: DEBUG nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.075120] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d26e59b-8dce-4e01-9c0a-8028d1e1287e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.323259] env[62000]: DEBUG nova.network.neutron [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updated VIF entry in instance network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1049.323679] env[62000]: DEBUG nova.network.neutron [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.433820] env[62000]: DEBUG nova.objects.instance [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 0246b032-505e-4bc5-bfc0-5779ff564626 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1049.453284] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1049.506090] env[62000]: DEBUG nova.scheduler.client.report [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1049.591836] env[62000]: INFO nova.compute.manager [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Took 16.34 seconds to build instance. [ 1049.782878] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1049.783180] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1049.783427] env[62000]: INFO nova.compute.manager [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attaching volume 9a5ba056-cc2d-45ea-aef7-f354937aa66f to /dev/sdb [ 1049.813459] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbfb0a84-591a-4099-885f-30ede5f19733 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.820014] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f780daeb-0468-42f4-9b07-0af7fc84b50e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.826168] env[62000]: DEBUG oslo_concurrency.lockutils [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.826456] env[62000]: DEBUG nova.compute.manager [req-5ab086c8-7614-46d8-9a8c-59c70c0fc6d5 req-b8890802-ee25-4a4c-82b4-08d01ccdfec9 service nova] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Received event network-vif-deleted-40a0a660-2c5f-4d14-829e-1c85ff3fc2bb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1049.832671] env[62000]: DEBUG nova.virt.block_device [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating existing volume attachment record: 2fc113c4-8f33-4900-9172-6ec6a9289ed3 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1049.955682] env[62000]: INFO nova.compute.manager [-] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 1.02 seconds to deallocate network for instance. [ 1049.967726] env[62000]: WARNING nova.volume.cinder [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Attachment ac743b5f-de32-45c3-ac43-579d33e281a5 does not exist. Ignoring.: cinderclient.exceptions.NotFound: Volume attachment could not be found with filter: attachment_id = ac743b5f-de32-45c3-ac43-579d33e281a5. (HTTP 404) (Request-ID: req-a0b9df1b-b647-4fdf-b907-7637c5b97ea9) [ 1049.968015] env[62000]: INFO nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Took 0.01 seconds to detach 1 volumes for instance. [ 1049.969971] env[62000]: DEBUG nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Deleting volume: d6f79cc7-55cc-4d38-9301-08c8d382294f {{(pid=62000) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1049.983737] env[62000]: WARNING nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Failed to delete volume: d6f79cc7-55cc-4d38-9301-08c8d382294f due to Volume d6f79cc7-55cc-4d38-9301-08c8d382294f could not be found.: nova.exception.VolumeNotFound: Volume d6f79cc7-55cc-4d38-9301-08c8d382294f could not be found. [ 1050.010620] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.719s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.013686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.620s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1050.013920] env[62000]: DEBUG nova.objects.instance [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'resources' on Instance uuid f8f26be0-f7d1-43e9-886d-c19d385e6935 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1050.036032] env[62000]: INFO nova.scheduler.client.report [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance 67af85f1-9134-440d-a5df-09ec7d3e72a0 [ 1050.093599] env[62000]: DEBUG oslo_concurrency.lockutils [None req-9dbc50fd-0727-46ee-bee3-51cab55d06fb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.846s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.440801] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8b12da68-6a5b-48c1-b3fc-0dd96bb0c816 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.236s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.489459] env[62000]: INFO nova.compute.manager [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance disappeared during terminate [ 1050.489703] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d466eed7-6b17-47b1-b1bd-64558c9701ab tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "9ee49e0b-6d37-4826-8d8a-bfb3752af4f8" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 1.667s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.544605] env[62000]: DEBUG oslo_concurrency.lockutils [None req-29395ab6-aa54-4226-9203-47a12f8b8343 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "67af85f1-9134-440d-a5df-09ec7d3e72a0" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.340s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1050.633323] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a55e867b-baed-42d9-89d0-fe2c93b8c28f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.641275] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3bff6fc-ddb4-4491-9839-75052aaa167b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.672369] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942cffc5-1b6d-4463-b62d-7f0323fddcac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.679313] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6c2d1b-6fba-41b1-864d-c8c4831d34da {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.692910] env[62000]: DEBUG nova.compute.provider_tree [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.195626] env[62000]: DEBUG nova.scheduler.client.report [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.234429] env[62000]: DEBUG nova.compute.manager [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Stashing vm_state: active {{(pid=62000) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1051.295269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.295380] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.295547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.296689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.296689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.297998] env[62000]: INFO nova.compute.manager [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Terminating instance [ 1051.301478] env[62000]: DEBUG nova.compute.manager [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.301672] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.302587] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d82d630-669c-4d3c-9401-63b0cef8fcd4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.310237] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.310464] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7bdd0613-6816-4aa0-981b-c1824685f0d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.316199] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1051.316199] env[62000]: value = "task-882880" [ 1051.316199] env[62000]: _type = "Task" [ 1051.316199] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.323798] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882880, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.498047] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.498269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.498496] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.498686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.498863] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.501276] env[62000]: INFO nova.compute.manager [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Terminating instance [ 1051.503215] env[62000]: DEBUG nova.compute.manager [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.503431] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1051.504260] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf449449-ff9a-46fc-8445-e2d1541abf16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.511705] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1051.511943] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3f0bed8a-06ba-4d96-9137-a097257e9bee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.518100] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1051.518100] env[62000]: value = "task-882881" [ 1051.518100] env[62000]: _type = "Task" [ 1051.518100] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.525621] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882881, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.700949] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.687s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.724676] env[62000]: INFO nova.scheduler.client.report [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted allocations for instance f8f26be0-f7d1-43e9-886d-c19d385e6935 [ 1051.751130] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.751471] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.829081] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882880, 'name': PowerOffVM_Task, 'duration_secs': 0.175953} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.829578] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1051.829903] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1051.830301] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e2898722-1f78-466d-9ee5-0b0989685f1f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.896434] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1051.896684] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1051.896969] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleting the datastore file [datastore1] 44a54bd7-a35c-49ed-85ed-346830cee6ad {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.897399] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4f9b0a9e-1252-475d-b4ec-e8ff05e26c41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.905398] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for the task: (returnval){ [ 1051.905398] env[62000]: value = "task-882883" [ 1051.905398] env[62000]: _type = "Task" [ 1051.905398] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.912581] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882883, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.029661] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882881, 'name': PowerOffVM_Task, 'duration_secs': 0.214014} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.029661] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1052.029661] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1052.029661] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-be145227-de95-422d-88b0-c8aa36cb3f3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.098752] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1052.098984] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1052.099198] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleting the datastore file [datastore2] 0246b032-505e-4bc5-bfc0-5779ff564626 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1052.099475] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e8c9d7ad-09db-4311-857e-6681806eb5db {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.105688] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1052.105688] env[62000]: value = "task-882886" [ 1052.105688] env[62000]: _type = "Task" [ 1052.105688] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.114300] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882886, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.233849] env[62000]: DEBUG oslo_concurrency.lockutils [None req-85bf2150-0c87-44d7-b632-320942b36754 tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f8f26be0-f7d1-43e9-886d-c19d385e6935" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.856s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.256087] env[62000]: INFO nova.compute.claims [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1052.415052] env[62000]: DEBUG oslo_vmware.api [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Task: {'id': task-882883, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.134326} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.415394] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.415596] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.415786] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.415964] env[62000]: INFO nova.compute.manager [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1052.416240] env[62000]: DEBUG oslo.service.loopingcall [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.416465] env[62000]: DEBUG nova.compute.manager [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.416540] env[62000]: DEBUG nova.network.neutron [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.617137] env[62000]: DEBUG oslo_vmware.api [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882886, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.176512} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.617420] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.617616] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1052.617800] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1052.617976] env[62000]: INFO nova.compute.manager [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1052.618314] env[62000]: DEBUG oslo.service.loopingcall [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.618551] env[62000]: DEBUG nova.compute.manager [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.618650] env[62000]: DEBUG nova.network.neutron [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1052.713827] env[62000]: DEBUG nova.compute.manager [req-139a2e69-0328-4891-8b2d-3eab8fda4a08 req-cbeed835-1b33-4054-a7ab-e6cc469f5f75 service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Received event network-vif-deleted-d42587d8-a09e-4869-b4d4-6f8265136d23 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1052.714064] env[62000]: INFO nova.compute.manager [req-139a2e69-0328-4891-8b2d-3eab8fda4a08 req-cbeed835-1b33-4054-a7ab-e6cc469f5f75 service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Neutron deleted interface d42587d8-a09e-4869-b4d4-6f8265136d23; detaching it from the instance and deleting it from the info cache [ 1052.714363] env[62000]: DEBUG nova.network.neutron [req-139a2e69-0328-4891-8b2d-3eab8fda4a08 req-cbeed835-1b33-4054-a7ab-e6cc469f5f75 service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.763907] env[62000]: INFO nova.compute.resource_tracker [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating resource usage from migration 857285ec-f4d0-4fa0-a980-a9467e9444f1 [ 1052.873023] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd9ec80-6ef3-4f9e-99e5-095ca86e0439 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.880441] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc736d7b-ed05-407e-aee0-7c089262262b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.910361] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79e1d645-6a12-4f23-8858-fca636d93d15 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.917857] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62ddc0a2-c3de-42eb-b6df-ea4d04b1e3d5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.933613] env[62000]: DEBUG nova.compute.provider_tree [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.187812] env[62000]: DEBUG nova.network.neutron [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.217341] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-408d07f9-28b2-40f8-a295-cc9cce81c4bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.226829] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f861ddf4-9246-45db-9b29-57611276fcf8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.255266] env[62000]: DEBUG nova.compute.manager [req-139a2e69-0328-4891-8b2d-3eab8fda4a08 req-cbeed835-1b33-4054-a7ab-e6cc469f5f75 service nova] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Detach interface failed, port_id=d42587d8-a09e-4869-b4d4-6f8265136d23, reason: Instance 44a54bd7-a35c-49ed-85ed-346830cee6ad could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1053.437068] env[62000]: DEBUG nova.scheduler.client.report [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1053.571989] env[62000]: DEBUG nova.network.neutron [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.691401] env[62000]: INFO nova.compute.manager [-] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Took 1.27 seconds to deallocate network for instance. [ 1053.941837] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.190s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.942167] env[62000]: INFO nova.compute.manager [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Migrating [ 1054.077385] env[62000]: INFO nova.compute.manager [-] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Took 1.46 seconds to deallocate network for instance. [ 1054.198373] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.198708] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.198950] env[62000]: DEBUG nova.objects.instance [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lazy-loading 'resources' on Instance uuid 44a54bd7-a35c-49ed-85ed-346830cee6ad {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.376980] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1054.377346] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201649', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'name': 'volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'serial': '9a5ba056-cc2d-45ea-aef7-f354937aa66f'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1054.378286] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c1eda2f-f5c3-446f-91b1-69e47190b762 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.394359] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ff7379-7adc-49a3-9cc4-fc3023f8d221 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.418600] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f/volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1054.418861] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-592eec5e-cdb1-4fdd-9c70-eeb1cff49728 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.438652] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1054.438652] env[62000]: value = "task-882888" [ 1054.438652] env[62000]: _type = "Task" [ 1054.438652] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.451220] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882888, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.458709] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.458709] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.458709] env[62000]: DEBUG nova.network.neutron [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1054.583196] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.744467] env[62000]: DEBUG nova.compute.manager [req-a7679d4b-93c7-4bf7-8809-5515019368f5 req-06cf4312-1cf0-426d-b437-22223f0796e5 service nova] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Received event network-vif-deleted-40d67815-d1be-4bab-8b61-bf89d3d62589 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1054.805377] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dacdaa-28e1-48d7-a4b3-0d062e41abda {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.813062] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb4bc07-6c0e-49a3-8b69-907ee68c3291 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.841604] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcd0725-ab48-4123-82c1-fb5f88873fd8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.848350] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9ed4437-c7ef-4e26-8850-32f26771b8b7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.860854] env[62000]: DEBUG nova.compute.provider_tree [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1054.948732] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882888, 'name': ReconfigVM_Task, 'duration_secs': 0.339385} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.949090] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f/volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1054.953891] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f840c05-d4e6-4149-ba8b-5279bf675581 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.970874] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1054.970874] env[62000]: value = "task-882889" [ 1054.970874] env[62000]: _type = "Task" [ 1054.970874] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.978690] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.161347] env[62000]: DEBUG nova.network.neutron [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.363858] env[62000]: DEBUG nova.scheduler.client.report [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.480935] env[62000]: DEBUG oslo_vmware.api [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882889, 'name': ReconfigVM_Task, 'duration_secs': 0.137909} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.480935] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201649', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'name': 'volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'serial': '9a5ba056-cc2d-45ea-aef7-f354937aa66f'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1055.664246] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.868648] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.670s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.871376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.288s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1055.871708] env[62000]: DEBUG nova.objects.instance [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'resources' on Instance uuid 0246b032-505e-4bc5-bfc0-5779ff564626 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1055.889231] env[62000]: INFO nova.scheduler.client.report [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Deleted allocations for instance 44a54bd7-a35c-49ed-85ed-346830cee6ad [ 1056.230132] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.230420] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.230640] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.230825] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.230994] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.233062] env[62000]: INFO nova.compute.manager [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Terminating instance [ 1056.234754] env[62000]: DEBUG nova.compute.manager [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.234955] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1056.235792] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31eef74-3206-400d-af35-237e9ee220c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.243507] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1056.243698] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-551c3429-69f3-42fc-8266-72d8409018a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.249643] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1056.249643] env[62000]: value = "task-882890" [ 1056.249643] env[62000]: _type = "Task" [ 1056.249643] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.257019] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882890, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.397348] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a55c9e6c-4965-49de-90e6-c2e59ca6bd06 tempest-ServersTestJSON-928621739 tempest-ServersTestJSON-928621739-project-member] Lock "44a54bd7-a35c-49ed-85ed-346830cee6ad" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.102s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.481649] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45a26f71-d4e0-43ab-ac06-d2154fe26a5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.489288] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032326da-f4b0-464c-b712-a0a28876dcad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.527024] env[62000]: DEBUG nova.objects.instance [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1056.533119] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3386aa-31bf-4b42-8842-4d3124915d2b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.538840] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0822bfb5-ccfb-46c6-a9da-bbdfa4d0a585 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.553290] env[62000]: DEBUG nova.compute.provider_tree [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1056.761922] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882890, 'name': PowerOffVM_Task, 'duration_secs': 0.204509} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.762280] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1056.762455] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1056.762704] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8bd8ac1-2011-48e6-8003-0e63aa32c728 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.828768] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1056.829028] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1056.829225] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleting the datastore file [datastore1] f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.829496] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-45da90d0-22ba-4149-b9c6-983d2b7f7782 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.836104] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for the task: (returnval){ [ 1056.836104] env[62000]: value = "task-882892" [ 1056.836104] env[62000]: _type = "Task" [ 1056.836104] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.843646] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882892, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.032676] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3910093a-81a4-4ffe-9d0d-222ee5a89fa6 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.249s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.055921] env[62000]: DEBUG nova.scheduler.client.report [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1057.178888] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49f566c8-4faa-496b-b598-68a3e5cc2483 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.198683] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 0 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1057.347886] env[62000]: DEBUG oslo_vmware.api [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Task: {'id': task-882892, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.127386} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.348481] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.348481] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1057.348481] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1057.348661] env[62000]: INFO nova.compute.manager [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1057.348896] env[62000]: DEBUG oslo.service.loopingcall [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.349104] env[62000]: DEBUG nova.compute.manager [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.349201] env[62000]: DEBUG nova.network.neutron [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1057.560563] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.689s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.596300] env[62000]: INFO nova.scheduler.client.report [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted allocations for instance 0246b032-505e-4bc5-bfc0-5779ff564626 [ 1057.630477] env[62000]: DEBUG nova.compute.manager [req-3ef3c31d-b697-4197-aaa2-7544094cba3b req-1d9bd89f-2c5f-480a-9ccc-ee75b26645e4 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Received event network-vif-deleted-adbea477-8af6-4cb5-82e6-1292553013f2 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1057.630477] env[62000]: INFO nova.compute.manager [req-3ef3c31d-b697-4197-aaa2-7544094cba3b req-1d9bd89f-2c5f-480a-9ccc-ee75b26645e4 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Neutron deleted interface adbea477-8af6-4cb5-82e6-1292553013f2; detaching it from the instance and deleting it from the info cache [ 1057.630622] env[62000]: DEBUG nova.network.neutron [req-3ef3c31d-b697-4197-aaa2-7544094cba3b req-1d9bd89f-2c5f-480a-9ccc-ee75b26645e4 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1057.705026] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1057.705369] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-324bfed6-1a76-49df-9385-781c0c2bceeb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.712113] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1057.712113] env[62000]: value = "task-882893" [ 1057.712113] env[62000]: _type = "Task" [ 1057.712113] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.721261] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.887723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.887723] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.105099] env[62000]: DEBUG nova.network.neutron [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.106434] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d11d27a9-52f0-4a26-9a68-72e68a099b7f tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "0246b032-505e-4bc5-bfc0-5779ff564626" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.608s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.133238] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ce8729f7-8e82-4861-9fa3-95c499389a8d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.142536] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dbee850-c5c2-4cf4-9bfc-a1d9d3c01961 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.170910] env[62000]: DEBUG nova.compute.manager [req-3ef3c31d-b697-4197-aaa2-7544094cba3b req-1d9bd89f-2c5f-480a-9ccc-ee75b26645e4 service nova] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Detach interface failed, port_id=adbea477-8af6-4cb5-82e6-1292553013f2, reason: Instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1058.221234] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882893, 'name': PowerOffVM_Task, 'duration_secs': 0.326412} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.221535] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1058.221733] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 17 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1058.391573] env[62000]: DEBUG nova.compute.utils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.609386] env[62000]: INFO nova.compute.manager [-] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Took 1.26 seconds to deallocate network for instance. [ 1058.728707] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1058.728981] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1058.729177] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1058.729357] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1058.729514] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1058.729669] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1058.729882] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1058.730074] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1058.730274] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1058.730447] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1058.730625] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1058.735669] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47162917-60bc-4fcb-a81e-e370170c0a32 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.753241] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1058.753241] env[62000]: value = "task-882894" [ 1058.753241] env[62000]: _type = "Task" [ 1058.753241] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.764835] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882894, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.894374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1059.115194] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.115309] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.115667] env[62000]: DEBUG nova.objects.instance [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lazy-loading 'resources' on Instance uuid f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.263737] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882894, 'name': ReconfigVM_Task, 'duration_secs': 0.186235} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.264084] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 33 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1059.703568] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01f44479-8499-4f5d-9a37-b8d248603bb3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.711600] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2c60609-f9e9-4dda-8e55-630ff06ba77c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.741102] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3af9ec6-b87d-4c51-aa5c-2de5eef5425a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.748152] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a88288-cf87-46cb-8a32-487bf0705af7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.763736] env[62000]: DEBUG nova.compute.provider_tree [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.770236] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1059.770440] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1059.770599] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1059.770781] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1059.770924] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1059.771081] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1059.771285] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1059.771659] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1059.771659] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1059.771773] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1059.771923] env[62000]: DEBUG nova.virt.hardware [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1059.777145] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfiguring VM instance instance-00000069 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1059.777968] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8c13b30f-e602-4e1e-bfc3-5a3a5b4dd5c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.797254] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1059.797254] env[62000]: value = "task-882895" [ 1059.797254] env[62000]: _type = "Task" [ 1059.797254] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1059.805900] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882895, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.965835] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.966146] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.966398] env[62000]: INFO nova.compute.manager [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attaching volume 4c0b430f-8215-4544-a0fd-902ab32deb4c to /dev/sdc [ 1059.998899] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435d619e-2be3-4338-a37e-432f3815b2bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.005746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9604ca24-9b96-401c-abf9-e9f9cd370c0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.019649] env[62000]: DEBUG nova.virt.block_device [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating existing volume attachment record: 2590123b-351a-4e72-9e7f-b22333150950 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1060.266820] env[62000]: DEBUG nova.scheduler.client.report [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.307604] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882895, 'name': ReconfigVM_Task, 'duration_secs': 0.200179} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.307889] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfigured VM instance instance-00000069 to detach disk 2000 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1060.308671] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bdb7775-4ccc-483e-af1f-4a7f5a9562c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.330393] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1060.330706] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1b8510dc-dcf3-4855-97e6-3cff44936907 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.348605] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1060.348605] env[62000]: value = "task-882897" [ 1060.348605] env[62000]: _type = "Task" [ 1060.348605] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.357302] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882897, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.544542] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.544808] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.771686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.656s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.793174] env[62000]: INFO nova.scheduler.client.report [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Deleted allocations for instance f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c [ 1060.860892] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882897, 'name': ReconfigVM_Task, 'duration_secs': 0.328186} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.861325] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17/6d2eaac3-c707-4850-8fc7-9edd56c8bc17.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1060.861707] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 50 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.049734] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1061.300780] env[62000]: DEBUG oslo_concurrency.lockutils [None req-df9d8d70-c656-478d-b190-fde59b0c118f tempest-ServerActionsTestOtherA-80520463 tempest-ServerActionsTestOtherA-80520463-project-member] Lock "f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.070s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.369736] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f6da07-76fb-4e7e-aa56-37634a46f7b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.388659] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97a8e099-ed0c-4f75-84c8-5317fa309a8e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.405486] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 67 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1061.574032] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.574145] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1061.576154] env[62000]: INFO nova.compute.claims [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1061.971033] env[62000]: DEBUG nova.network.neutron [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Port 45862b3c-e63d-4af1-9192-270c4f66cabb binding to destination host cpu-1 is already ACTIVE {{(pid=62000) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3171}} [ 1062.699246] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8984a8cd-0521-4064-bed3-bab9dcd06a85 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.708942] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf861a4-5a5f-4cd4-b8bd-c93b00002a82 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.739849] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4e1bc94-937a-4983-b3a0-a2667833751b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.747794] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83a1a178-5b8a-4d90-a542-58ba34bf29c7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.760624] env[62000]: DEBUG nova.compute.provider_tree [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1062.888199] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.888513] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1062.888676] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1062.993497] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.993846] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.993907] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1063.282813] env[62000]: ERROR nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [req-9c104cd9-2bf5-44fd-bb38-14d82e00c466] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-9c104cd9-2bf5-44fd-bb38-14d82e00c466"}]} [ 1063.301101] env[62000]: DEBUG nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1063.320697] env[62000]: DEBUG nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1063.320945] env[62000]: DEBUG nova.compute.provider_tree [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1063.333225] env[62000]: DEBUG nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1063.352029] env[62000]: DEBUG nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1063.435322] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e150f881-5e91-4ec8-93d0-3d455da2358c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.442988] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d5c5b8-0c89-4c8d-9021-d048ed550b6a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.473396] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e05a1d08-b51e-4d9a-ab2b-73a3fe3f10c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.481025] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2920b96e-5c4a-41fa-a4d7-99e6f37471e8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.493991] env[62000]: DEBUG nova.compute.provider_tree [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1064.028262] env[62000]: DEBUG nova.scheduler.client.report [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 139 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1064.028583] env[62000]: DEBUG nova.compute.provider_tree [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 139 to 140 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1064.028744] env[62000]: DEBUG nova.compute.provider_tree [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1064.243599] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.243810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.244008] env[62000]: DEBUG nova.network.neutron [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1064.533429] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.959s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.533955] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1064.569108] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1064.569383] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201650', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'name': 'volume-4c0b430f-8215-4544-a0fd-902ab32deb4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'serial': '4c0b430f-8215-4544-a0fd-902ab32deb4c'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1064.570356] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db3de0ca-30a3-4c8c-a03b-9e967debbddb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.586919] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfdfe869-8249-4911-90ed-2af7c88914b8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.613714] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] volume-4c0b430f-8215-4544-a0fd-902ab32deb4c/volume-4c0b430f-8215-4544-a0fd-902ab32deb4c.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.613948] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d9f8ee25-a9bc-4a78-8b14-9033606f1a95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.631384] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1064.631384] env[62000]: value = "task-882899" [ 1064.631384] env[62000]: _type = "Task" [ 1064.631384] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.641225] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882899, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.940397] env[62000]: DEBUG nova.network.neutron [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.039146] env[62000]: DEBUG nova.compute.utils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.040618] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.040887] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1065.080036] env[62000]: DEBUG nova.policy [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8d0b65722357449ab64736e5e1eb713f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8ce3270692934d2c9c1330a45c0e059e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1065.140679] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882899, 'name': ReconfigVM_Task, 'duration_secs': 0.370031} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.140958] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfigured VM instance instance-00000062 to attach disk [datastore1] volume-4c0b430f-8215-4544-a0fd-902ab32deb4c/volume-4c0b430f-8215-4544-a0fd-902ab32deb4c.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1065.145614] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccc5e20c-0bd7-4327-8b80-5f4589a5aa88 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.159948] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1065.159948] env[62000]: value = "task-882900" [ 1065.159948] env[62000]: _type = "Task" [ 1065.159948] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.167353] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882900, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.358496] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Successfully created port: 679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1065.442822] env[62000]: DEBUG oslo_concurrency.lockutils [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.544026] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1065.668982] env[62000]: DEBUG oslo_vmware.api [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882900, 'name': ReconfigVM_Task, 'duration_secs': 0.130492} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.669317] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201650', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'name': 'volume-4c0b430f-8215-4544-a0fd-902ab32deb4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'serial': '4c0b430f-8215-4544-a0fd-902ab32deb4c'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1065.967736] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4129842f-4e49-45a5-93ce-0ed0612efd3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.987376] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5faa6abd-6c4f-4838-9ae0-4c21cc0ef40d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.993958] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 83 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1066.410937] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Didn't find any instances for network info cache update. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1066.411249] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.411377] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.411529] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.411681] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.411825] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.411969] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.412174] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1066.412333] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.500304] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1066.500616] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5dda1de3-ce81-4ede-b2ba-2dc339f6755e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.507834] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1066.507834] env[62000]: value = "task-882901" [ 1066.507834] env[62000]: _type = "Task" [ 1066.507834] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.518877] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882901, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.553486] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1066.579556] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1066.579806] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1066.579977] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1066.580184] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1066.580339] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1066.580491] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1066.580706] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1066.580870] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1066.581054] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1066.581229] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1066.581406] env[62000]: DEBUG nova.virt.hardware [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1066.582450] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-594a3816-7dfb-435c-af7f-d13c53b91347 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.590420] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda5cc22-08a6-4b34-b02e-65e55803d43a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.708100] env[62000]: DEBUG nova.objects.instance [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1066.827500] env[62000]: DEBUG nova.compute.manager [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Received event network-vif-plugged-679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1066.827660] env[62000]: DEBUG oslo_concurrency.lockutils [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.828309] env[62000]: DEBUG oslo_concurrency.lockutils [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.828572] env[62000]: DEBUG oslo_concurrency.lockutils [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.828828] env[62000]: DEBUG nova.compute.manager [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] No waiting events found dispatching network-vif-plugged-679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1066.828944] env[62000]: WARNING nova.compute.manager [req-cbbbfd7b-3540-4df8-b18b-a8cf8a39b7b6 req-a1daadcf-88f4-4479-a77c-49538cc88d84 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Received unexpected event network-vif-plugged-679d672f-6163-425c-b4d9-c74d7d7b2a3f for instance with vm_state building and task_state spawning. [ 1066.915021] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.915265] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.915438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.915596] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1066.916775] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f41d9b9-ce47-4bd4-84ed-78620b440e73 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.924270] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ada204e-37aa-4247-ade5-7f05ecd09e0a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.938533] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a51025f-85ed-493d-9f55-229973d060fb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.944549] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c2a58bb-2cd7-44e3-aba8-154f0d7a9956 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.972803] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180617MB free_disk=53GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1066.972963] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.973209] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.017410] env[62000]: DEBUG oslo_vmware.api [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882901, 'name': PowerOnVM_Task, 'duration_secs': 0.347411} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.017680] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1067.017869] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-5a516781-d205-4df4-b7fa-4713f8e81aeb tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance '6d2eaac3-c707-4850-8fc7-9edd56c8bc17' progress to 100 {{(pid=62000) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1354}} [ 1067.213128] env[62000]: DEBUG oslo_concurrency.lockutils [None req-8131a662-1d76-48b3-b14c-408efc02b8f2 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.247s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.309400] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Successfully updated port: 679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.336063] env[62000]: DEBUG nova.compute.manager [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Received event network-changed-679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1067.336251] env[62000]: DEBUG nova.compute.manager [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Refreshing instance network info cache due to event network-changed-679d672f-6163-425c-b4d9-c74d7d7b2a3f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1067.336662] env[62000]: DEBUG oslo_concurrency.lockutils [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] Acquiring lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.336878] env[62000]: DEBUG oslo_concurrency.lockutils [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] Acquired lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.337156] env[62000]: DEBUG nova.network.neutron [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Refreshing network info cache for port 679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1067.556589] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1067.556915] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1067.813240] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.868742] env[62000]: DEBUG nova.network.neutron [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1067.936814] env[62000]: DEBUG nova.network.neutron [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.980850] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Applying migration context for instance 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 as it has an incoming, in-progress migration 857285ec-f4d0-4fa0-a980-a9467e9444f1. Migration status is post-migrating {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1067.981795] env[62000]: INFO nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating resource usage from migration 857285ec-f4d0-4fa0-a980-a9467e9444f1 [ 1068.000395] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 9ccf07fa-90db-40b9-af5d-b47ee5773f9c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.000546] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 15fde663-fac4-4198-962f-8f814f1317f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.000675] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 19410eb7-0fec-4270-89da-04a2975fc050 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.000794] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Migration 857285ec-f4d0-4fa0-a980-a9467e9444f1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1068.000912] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.001038] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 58b4c1b2-11b2-4d52-91c6-6df943b500f3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1068.001219] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1068.001353] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1728MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1068.060801] env[62000]: INFO nova.compute.manager [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Detaching volume 9a5ba056-cc2d-45ea-aef7-f354937aa66f [ 1068.081736] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-671cde5a-be00-496a-a573-6fd00e810ea5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.089517] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be401060-8d8d-4732-8ab2-4c33acc55a9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.093790] env[62000]: INFO nova.virt.block_device [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attempting to driver detach volume 9a5ba056-cc2d-45ea-aef7-f354937aa66f from mountpoint /dev/sdb [ 1068.094035] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1068.094235] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201649', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'name': 'volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'serial': '9a5ba056-cc2d-45ea-aef7-f354937aa66f'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1068.095215] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-604251a1-8a88-4acc-b417-19d963fd5193 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.121554] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6533afe8-1af7-475e-814b-49e7f2c88be6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.141652] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4686da-28b1-4520-87e4-6b363b4e2e84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.146761] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2c432f8-6e29-4718-8f40-001eeeec9a65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.153995] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d976e28-877c-44e5-bad3-98552322d8ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.163241] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1068.184505] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc2a58e9-de2f-4550-8da7-951e1c8b27a0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.199427] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] The volume has not been displaced from its original location: [datastore1] volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f/volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1068.204659] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfiguring VM instance instance-00000062 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1068.205140] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-76a66e78-ef2e-4a2b-af9f-a32c82d14b06 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.223712] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1068.223712] env[62000]: value = "task-882902" [ 1068.223712] env[62000]: _type = "Task" [ 1068.223712] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.231367] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882902, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.439644] env[62000]: DEBUG oslo_concurrency.lockutils [req-37d0289f-cfbd-47a2-b9d2-52299b54cc36 req-31772660-3126-465e-b939-81123f19ae1a service nova] Releasing lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.440010] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.440176] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1068.666381] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.733667] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882902, 'name': ReconfigVM_Task, 'duration_secs': 0.217576} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.733942] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfigured VM instance instance-00000062 to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1068.738526] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e2b0cf41-35c4-4ecc-9e8a-8146b18a2f6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.753694] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1068.753694] env[62000]: value = "task-882903" [ 1068.753694] env[62000]: _type = "Task" [ 1068.753694] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.761076] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.939660] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.940077] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.940353] env[62000]: DEBUG nova.compute.manager [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Going to confirm migration 7 {{(pid=62000) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1068.971700] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1069.104616] env[62000]: DEBUG nova.network.neutron [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating instance_info_cache with network_info: [{"id": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "address": "fa:16:3e:b4:ee:ff", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap679d672f-61", "ovs_interfaceid": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.171396] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1069.171704] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.198s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.264299] env[62000]: DEBUG oslo_vmware.api [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882903, 'name': ReconfigVM_Task, 'duration_secs': 0.12891} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.264645] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201649', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'name': 'volume-9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '9a5ba056-cc2d-45ea-aef7-f354937aa66f', 'serial': '9a5ba056-cc2d-45ea-aef7-f354937aa66f'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1069.479350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.479591] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquired lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.479783] env[62000]: DEBUG nova.network.neutron [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1069.479982] env[62000]: DEBUG nova.objects.instance [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'info_cache' on Instance uuid 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1069.607188] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.607528] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Instance network_info: |[{"id": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "address": "fa:16:3e:b4:ee:ff", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap679d672f-61", "ovs_interfaceid": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1069.607955] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b4:ee:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cbe1725d-6711-4e92-9a4e-d4802651e7d0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '679d672f-6163-425c-b4d9-c74d7d7b2a3f', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1069.615412] env[62000]: DEBUG oslo.service.loopingcall [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1069.615618] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1069.615843] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52b271ef-bcb5-4308-93a7-ad077f7811b5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.635776] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1069.635776] env[62000]: value = "task-882904" [ 1069.635776] env[62000]: _type = "Task" [ 1069.635776] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.642863] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882904, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.806835] env[62000]: DEBUG nova.objects.instance [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1070.146136] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882904, 'name': CreateVM_Task, 'duration_secs': 0.397498} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.146292] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1070.146954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.147144] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.147464] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1070.147709] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89505265-9d82-40de-b1d9-cb263ce5c3b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.151939] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1070.151939] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d7a5e5-a417-edaa-7735-fd6a85cc411a" [ 1070.151939] env[62000]: _type = "Task" [ 1070.151939] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.159581] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d7a5e5-a417-edaa-7735-fd6a85cc411a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.664748] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52d7a5e5-a417-edaa-7735-fd6a85cc411a, 'name': SearchDatastore_Task, 'duration_secs': 0.01883} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.666017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.666017] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1070.666017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1070.666017] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.666017] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1070.666247] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4129edf5-a385-4463-9548-661727a5dea6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.674294] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1070.674475] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1070.675174] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-47f0f669-246a-4259-9912-adf932e455d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.679885] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1070.679885] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c49f6e-95c6-2e4a-ca43-2fd01adf4ea9" [ 1070.679885] env[62000]: _type = "Task" [ 1070.679885] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.686731] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c49f6e-95c6-2e4a-ca43-2fd01adf4ea9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.688978] env[62000]: DEBUG nova.network.neutron [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [{"id": "45862b3c-e63d-4af1-9192-270c4f66cabb", "address": "fa:16:3e:fb:b2:f5", "network": {"id": "4d2d52b4-3a96-4273-94d3-a1018ef5c2a7", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-723753462-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "43e5c47870584d05abaf9de72d45cce2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "68ec9c06-8680-4a41-abad-cddbd1f768c9", "external-id": "nsx-vlan-transportzone-883", "segmentation_id": 883, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap45862b3c-e6", "ovs_interfaceid": "45862b3c-e63d-4af1-9192-270c4f66cabb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.795392] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1070.814416] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ced4804b-a686-451d-9d2a-fd84a5177356 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.257s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1070.815410] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.020s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.189914] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c49f6e-95c6-2e4a-ca43-2fd01adf4ea9, 'name': SearchDatastore_Task, 'duration_secs': 0.01039} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.190797] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fca12ea1-5ac4-4f15-b189-20cac73f5c1b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.193335] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Releasing lock "refresh_cache-6d2eaac3-c707-4850-8fc7-9edd56c8bc17" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.193591] env[62000]: DEBUG nova.objects.instance [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lazy-loading 'migration_context' on Instance uuid 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1071.197707] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1071.197707] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f6ab1-cbbf-436c-5709-ba98faaa05f8" [ 1071.197707] env[62000]: _type = "Task" [ 1071.197707] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.206088] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f6ab1-cbbf-436c-5709-ba98faaa05f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.256907] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.257171] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.257356] env[62000]: INFO nova.compute.manager [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Shelving [ 1071.320543] env[62000]: INFO nova.compute.manager [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Detaching volume 4c0b430f-8215-4544-a0fd-902ab32deb4c [ 1071.351777] env[62000]: INFO nova.virt.block_device [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Attempting to driver detach volume 4c0b430f-8215-4544-a0fd-902ab32deb4c from mountpoint /dev/sdc [ 1071.352050] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1071.352284] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201650', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'name': 'volume-4c0b430f-8215-4544-a0fd-902ab32deb4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'serial': '4c0b430f-8215-4544-a0fd-902ab32deb4c'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1071.353182] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cd628d2-c3c7-4805-9554-e1ead1da0744 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.373532] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76577678-eff7-4de2-a97a-bfa1d7a5ef65 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.379517] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f90682-0c10-4828-859c-70912d5903d3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.398284] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb12d21b-ec1b-4679-a7b6-11033a3442f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.412246] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] The volume has not been displaced from its original location: [datastore1] volume-4c0b430f-8215-4544-a0fd-902ab32deb4c/volume-4c0b430f-8215-4544-a0fd-902ab32deb4c.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1071.417279] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfiguring VM instance instance-00000062 to detach disk 2002 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1071.417528] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b96db6f7-aa38-4500-b9a7-b9541bde8de2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.434181] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1071.434181] env[62000]: value = "task-882905" [ 1071.434181] env[62000]: _type = "Task" [ 1071.434181] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.442643] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882905, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.696812] env[62000]: DEBUG nova.objects.base [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Object Instance<6d2eaac3-c707-4850-8fc7-9edd56c8bc17> lazy-loaded attributes: info_cache,migration_context {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1071.697734] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dec3dc3e-5294-4429-ac63-14f4eeb1ac9c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.709018] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521f6ab1-cbbf-436c-5709-ba98faaa05f8, 'name': SearchDatastore_Task, 'duration_secs': 0.011806} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.721663] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.721936] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 58b4c1b2-11b2-4d52-91c6-6df943b500f3/58b4c1b2-11b2-4d52-91c6-6df943b500f3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1071.722778] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b177c2d8-2ce0-44a2-af0b-95c05bd7da90 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.724520] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b39cefe-0b04-4561-8025-7c9c9d0cc938 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.729731] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1071.729731] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529218b4-0875-4e2a-7f13-7dbb01a576fc" [ 1071.729731] env[62000]: _type = "Task" [ 1071.729731] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.733365] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1071.733365] env[62000]: value = "task-882906" [ 1071.733365] env[62000]: _type = "Task" [ 1071.733365] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.738966] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529218b4-0875-4e2a-7f13-7dbb01a576fc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.743106] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882906, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.763767] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1071.763947] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdb16bb4-3ad9-45f6-9453-f8370dc09682 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.769484] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1071.769484] env[62000]: value = "task-882907" [ 1071.769484] env[62000]: _type = "Task" [ 1071.769484] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.776582] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882907, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.943986] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882905, 'name': ReconfigVM_Task, 'duration_secs': 0.244803} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.943986] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Reconfigured VM instance instance-00000062 to detach disk 2002 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1071.948335] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1139c200-d4fd-4f69-bb74-acb0a2d1c6f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.964297] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1071.964297] env[62000]: value = "task-882908" [ 1071.964297] env[62000]: _type = "Task" [ 1071.964297] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.972689] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882908, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.242268] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882906, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.245713] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]529218b4-0875-4e2a-7f13-7dbb01a576fc, 'name': SearchDatastore_Task, 'duration_secs': 0.009078} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.246016] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1072.246269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1072.278543] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882907, 'name': PowerOffVM_Task, 'duration_secs': 0.158365} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.278805] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1072.279566] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bec1e9b-7cda-4939-9cac-9bb6501fcfbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.298331] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f95b55-25b8-4111-809f-8a5c38179adf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.473610] env[62000]: DEBUG oslo_vmware.api [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882908, 'name': ReconfigVM_Task, 'duration_secs': 0.169684} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.473925] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201650', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'name': 'volume-4c0b430f-8215-4544-a0fd-902ab32deb4c', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '9ccf07fa-90db-40b9-af5d-b47ee5773f9c', 'attached_at': '', 'detached_at': '', 'volume_id': '4c0b430f-8215-4544-a0fd-902ab32deb4c', 'serial': '4c0b430f-8215-4544-a0fd-902ab32deb4c'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1072.743471] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882906, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.607137} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.743731] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 58b4c1b2-11b2-4d52-91c6-6df943b500f3/58b4c1b2-11b2-4d52-91c6-6df943b500f3.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1072.743951] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1072.744224] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d399084-af05-4627-9cf1-b3ef3d546ba4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.751688] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1072.751688] env[62000]: value = "task-882909" [ 1072.751688] env[62000]: _type = "Task" [ 1072.751688] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.759088] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882909, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.807632] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1072.807956] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-3b3c7db2-93b7-4c3b-ab2a-e245a15f29ed {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.818015] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1072.818015] env[62000]: value = "task-882910" [ 1072.818015] env[62000]: _type = "Task" [ 1072.818015] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.825795] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882910, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.853283] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d154d1ab-9b3a-4e23-b664-cf87c37d96ee {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.860596] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ebf361b-2b60-45d1-a4fd-07c89ef803a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.889670] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac88b72-758c-491a-bf44-fe9bd3d2c112 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.896417] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd5e1e4-ff3a-430a-a68b-fb839e0506a2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.909073] env[62000]: DEBUG nova.compute.provider_tree [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1073.014595] env[62000]: DEBUG nova.objects.instance [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'flavor' on Instance uuid 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1073.261696] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882909, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.14326} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.263098] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1073.263098] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec98dfdf-20f3-42ad-84b2-3b63eb07c7f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.283718] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] 58b4c1b2-11b2-4d52-91c6-6df943b500f3/58b4c1b2-11b2-4d52-91c6-6df943b500f3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1073.283976] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-26435212-2462-447b-ba6f-e21a3572c921 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.302912] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1073.302912] env[62000]: value = "task-882911" [ 1073.302912] env[62000]: _type = "Task" [ 1073.302912] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.310390] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882911, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.325805] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882910, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.440321] env[62000]: DEBUG nova.scheduler.client.report [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 140 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1073.440789] env[62000]: DEBUG nova.compute.provider_tree [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 140 to 141 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1073.441033] env[62000]: DEBUG nova.compute.provider_tree [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1073.812905] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882911, 'name': ReconfigVM_Task, 'duration_secs': 0.301334} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.813256] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfigured VM instance instance-0000006a to attach disk [datastore1] 58b4c1b2-11b2-4d52-91c6-6df943b500f3/58b4c1b2-11b2-4d52-91c6-6df943b500f3.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1073.813887] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6af4f050-fede-4bcf-8fd8-a01b127a9432 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.819620] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1073.819620] env[62000]: value = "task-882912" [ 1073.819620] env[62000]: _type = "Task" [ 1073.819620] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.829753] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882910, 'name': CreateSnapshot_Task, 'duration_secs': 0.861062} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.832760] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1073.833010] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882912, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.833704] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942e03ab-e0b8-475a-8df7-3dee78ab7ab0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.022729] env[62000]: DEBUG oslo_concurrency.lockutils [None req-2dbca4b8-7736-4ca5-9457-f1e7f62dbbd0 tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.207s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.331538] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882912, 'name': Rename_Task, 'duration_secs': 0.134081} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.331879] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1074.332202] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1dc0cd7b-9e22-4769-b037-3cfa22c25ff2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.339219] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1074.339219] env[62000]: value = "task-882913" [ 1074.339219] env[62000]: _type = "Task" [ 1074.339219] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.351199] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1074.351485] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.351699] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-5ef07929-0a09-444d-b577-5d8cd3d6f739 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.359605] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1074.359605] env[62000]: value = "task-882914" [ 1074.359605] env[62000]: _type = "Task" [ 1074.359605] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.368547] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882914, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.451413] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.205s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.848383] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882913, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.869934] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882914, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.005488] env[62000]: INFO nova.scheduler.client.report [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocation for migration 857285ec-f4d0-4fa0-a980-a9467e9444f1 [ 1075.190692] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.191158] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.191238] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.191443] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.191628] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.193943] env[62000]: INFO nova.compute.manager [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Terminating instance [ 1075.195725] env[62000]: DEBUG nova.compute.manager [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1075.195929] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1075.196770] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d195876-6268-4c12-8858-5cf4499cc24a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.204293] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1075.204528] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d2063f8e-399a-4e73-8d57-48b53918bea1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.210612] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1075.210612] env[62000]: value = "task-882915" [ 1075.210612] env[62000]: _type = "Task" [ 1075.210612] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.218382] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.348247] env[62000]: DEBUG oslo_vmware.api [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882913, 'name': PowerOnVM_Task, 'duration_secs': 0.702028} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.348623] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1075.348850] env[62000]: INFO nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Took 8.80 seconds to spawn the instance on the hypervisor. [ 1075.349048] env[62000]: DEBUG nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.349785] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0225727-12cf-4309-902f-b5c191958974 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.369492] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882914, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.511372] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.571s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.720583] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882915, 'name': PowerOffVM_Task, 'duration_secs': 0.282839} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.720831] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1075.720974] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1075.721269] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b68ed353-f556-4e58-84ff-788d6958d0f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.788367] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1075.788599] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1075.788796] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleting the datastore file [datastore2] 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1075.789063] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8b4f4cfe-2734-4c2f-87db-c64832dbdaba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.795566] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for the task: (returnval){ [ 1075.795566] env[62000]: value = "task-882917" [ 1075.795566] env[62000]: _type = "Task" [ 1075.795566] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.803066] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.869260] env[62000]: INFO nova.compute.manager [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Took 14.31 seconds to build instance. [ 1075.875517] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882914, 'name': CloneVM_Task, 'duration_secs': 1.300454} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.875517] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Created linked-clone VM from snapshot [ 1075.875982] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-762bdb8b-aa36-43d6-ac65-4233188d36a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.882988] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Uploading image 38a40578-3968-46a5-b445-a5b08f645b60 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1075.916874] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1075.916874] env[62000]: value = "vm-201653" [ 1075.916874] env[62000]: _type = "VirtualMachine" [ 1075.916874] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1075.917204] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f0ce3b69-a448-4d30-a8ea-c70076469b4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.923434] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease: (returnval){ [ 1075.923434] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52da2647-9ae5-eb83-1698-d2c501cc293e" [ 1075.923434] env[62000]: _type = "HttpNfcLease" [ 1075.923434] env[62000]: } obtained for exporting VM: (result){ [ 1075.923434] env[62000]: value = "vm-201653" [ 1075.923434] env[62000]: _type = "VirtualMachine" [ 1075.923434] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1075.923681] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the lease: (returnval){ [ 1075.923681] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52da2647-9ae5-eb83-1698-d2c501cc293e" [ 1075.923681] env[62000]: _type = "HttpNfcLease" [ 1075.923681] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1075.929374] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1075.929374] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52da2647-9ae5-eb83-1698-d2c501cc293e" [ 1075.929374] env[62000]: _type = "HttpNfcLease" [ 1075.929374] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1076.305282] env[62000]: DEBUG oslo_vmware.api [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Task: {'id': task-882917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157002} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.305660] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1076.305766] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1076.305918] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1076.306113] env[62000]: INFO nova.compute.manager [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1076.306370] env[62000]: DEBUG oslo.service.loopingcall [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1076.306571] env[62000]: DEBUG nova.compute.manager [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1076.306668] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1076.371580] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ffb84182-24aa-41b5-93ec-d1fc45a088e0 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.827s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.433022] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1076.433022] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52da2647-9ae5-eb83-1698-d2c501cc293e" [ 1076.433022] env[62000]: _type = "HttpNfcLease" [ 1076.433022] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1076.433022] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1076.433022] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52da2647-9ae5-eb83-1698-d2c501cc293e" [ 1076.433022] env[62000]: _type = "HttpNfcLease" [ 1076.433022] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1076.433022] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d7aa217-71fb-445f-b4c5-06a901f39e4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.441307] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1076.441307] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1076.533757] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ba19eb58-26fa-4b9b-9a34-0d6e2589dd17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.893750] env[62000]: DEBUG nova.compute.manager [req-ac9d2c16-c821-470d-8e52-a68ff4ea7b24 req-6c4848a1-2a7b-4b50-a812-de22f11cbbcb service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Received event network-vif-deleted-5f0f0c25-4143-4468-b1e5-48897bd334ac {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1076.893981] env[62000]: INFO nova.compute.manager [req-ac9d2c16-c821-470d-8e52-a68ff4ea7b24 req-6c4848a1-2a7b-4b50-a812-de22f11cbbcb service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Neutron deleted interface 5f0f0c25-4143-4468-b1e5-48897bd334ac; detaching it from the instance and deleting it from the info cache [ 1076.894200] env[62000]: DEBUG nova.network.neutron [req-ac9d2c16-c821-470d-8e52-a68ff4ea7b24 req-6c4848a1-2a7b-4b50-a812-de22f11cbbcb service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.007829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.007829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.007829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.007829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.007829] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.010754] env[62000]: INFO nova.compute.manager [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Terminating instance [ 1077.012912] env[62000]: DEBUG nova.compute.manager [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1077.013290] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1077.016105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd4b1c75-f61c-40e8-9f60-aef0038b8495 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.022719] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1077.023202] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62f8cc74-2280-49da-9008-cad98bc4b113 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.030590] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1077.032144] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Cleaning up deleted instances with incomplete migration {{(pid=62000) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11264}} [ 1077.032144] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1077.032144] env[62000]: value = "task-882919" [ 1077.032144] env[62000]: _type = "Task" [ 1077.032144] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.042844] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882919, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.133394] env[62000]: DEBUG nova.compute.manager [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Received event network-changed-679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1077.133708] env[62000]: DEBUG nova.compute.manager [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Refreshing instance network info cache due to event network-changed-679d672f-6163-425c-b4d9-c74d7d7b2a3f. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1077.134440] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] Acquiring lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.135225] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] Acquired lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.136704] env[62000]: DEBUG nova.network.neutron [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Refreshing network info cache for port 679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1077.368509] env[62000]: DEBUG nova.network.neutron [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1077.398101] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9383a49a-bcea-4ce4-96a8-d1f2c9feabd5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.408076] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d53773e7-d067-48fb-912f-16263261ee50 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.442194] env[62000]: DEBUG nova.compute.manager [req-ac9d2c16-c821-470d-8e52-a68ff4ea7b24 req-6c4848a1-2a7b-4b50-a812-de22f11cbbcb service nova] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Detach interface failed, port_id=5f0f0c25-4143-4468-b1e5-48897bd334ac, reason: Instance 9ccf07fa-90db-40b9-af5d-b47ee5773f9c could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1077.543874] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882919, 'name': PowerOffVM_Task, 'duration_secs': 0.225927} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.544366] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1077.544697] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1077.545091] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e31f0361-5636-4aa1-9c8b-4b013033a88c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.612346] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1077.612686] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1077.612942] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleting the datastore file [datastore1] 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1077.613292] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5f2f7ec2-5ee2-47a0-bc99-13076e709c8d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.620699] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for the task: (returnval){ [ 1077.620699] env[62000]: value = "task-882921" [ 1077.620699] env[62000]: _type = "Task" [ 1077.620699] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.629375] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.870800] env[62000]: INFO nova.compute.manager [-] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Took 1.56 seconds to deallocate network for instance. [ 1077.877863] env[62000]: DEBUG nova.network.neutron [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updated VIF entry in instance network info cache for port 679d672f-6163-425c-b4d9-c74d7d7b2a3f. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1077.878412] env[62000]: DEBUG nova.network.neutron [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating instance_info_cache with network_info: [{"id": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "address": "fa:16:3e:b4:ee:ff", "network": {"id": "88f21116-cbe9-4a2a-a7cf-f6b01d58e030", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1033600554-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.153", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8ce3270692934d2c9c1330a45c0e059e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cbe1725d-6711-4e92-9a4e-d4802651e7d0", "external-id": "nsx-vlan-transportzone-679", "segmentation_id": 679, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap679d672f-61", "ovs_interfaceid": "679d672f-6163-425c-b4d9-c74d7d7b2a3f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.030542] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.131105] env[62000]: DEBUG oslo_vmware.api [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Task: {'id': task-882921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.225632} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.131426] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1078.131677] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1078.131885] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1078.132179] env[62000]: INFO nova.compute.manager [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1078.132481] env[62000]: DEBUG oslo.service.loopingcall [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1078.132732] env[62000]: DEBUG nova.compute.manager [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1078.132858] env[62000]: DEBUG nova.network.neutron [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1078.377276] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1078.377761] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.378045] env[62000]: DEBUG nova.objects.instance [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lazy-loading 'resources' on Instance uuid 9ccf07fa-90db-40b9-af5d-b47ee5773f9c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.380385] env[62000]: DEBUG oslo_concurrency.lockutils [req-a5594274-18ec-4694-b9d0-25af7b7e29ad req-163dea63-eed4-4a28-b968-02136802ae70 service nova] Releasing lock "refresh_cache-58b4c1b2-11b2-4d52-91c6-6df943b500f3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.850347] env[62000]: DEBUG nova.network.neutron [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.921787] env[62000]: DEBUG nova.compute.manager [req-d29a37a1-96c8-463e-8281-1d8e022bd4d4 req-e1684338-85cc-4144-8a87-6b7ac4eeba07 service nova] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Received event network-vif-deleted-45862b3c-e63d-4af1-9192-270c4f66cabb {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1078.964810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d86c182-b8ef-44f0-b5e8-93f6c3cd1e49 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.972397] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1e971e-4d8b-4f56-9755-54cefc6a3af4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.008892] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d180157d-bdb0-4a7a-92f8-b10152f565f6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.016764] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfddd59f-5d73-46a9-a778-f5c21fe5d53c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.030429] env[62000]: DEBUG nova.compute.provider_tree [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.353056] env[62000]: INFO nova.compute.manager [-] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Took 1.22 seconds to deallocate network for instance. [ 1079.534239] env[62000]: DEBUG nova.scheduler.client.report [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.860858] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.039767] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.662s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.042689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.182s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.042952] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.058655] env[62000]: INFO nova.scheduler.client.report [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Deleted allocations for instance 6d2eaac3-c707-4850-8fc7-9edd56c8bc17 [ 1080.060533] env[62000]: INFO nova.scheduler.client.report [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Deleted allocations for instance 9ccf07fa-90db-40b9-af5d-b47ee5773f9c [ 1080.570438] env[62000]: DEBUG oslo_concurrency.lockutils [None req-6c6b7a42-8c7f-47a3-91ac-399d31f9d8c0 tempest-DeleteServersTestJSON-807369734 tempest-DeleteServersTestJSON-807369734-project-member] Lock "6d2eaac3-c707-4850-8fc7-9edd56c8bc17" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.563s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.571920] env[62000]: DEBUG oslo_concurrency.lockutils [None req-19805f74-a254-4f43-84db-1bd1d725515b tempest-AttachVolumeTestJSON-2011993537 tempest-AttachVolumeTestJSON-2011993537-project-member] Lock "9ccf07fa-90db-40b9-af5d-b47ee5773f9c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.381s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1081.526893] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1081.527167] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.029660] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1083.030054] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1083.030054] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Rebuilding the list of instances to heal {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 1083.533755] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.533755] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.533755] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Forcefully refreshing network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 1083.533986] env[62000]: DEBUG nova.objects.instance [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lazy-loading 'info_cache' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.629812] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.630318] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.630388] env[62000]: INFO nova.compute.manager [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Shelving [ 1085.139684] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1085.139844] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2e68cd30-95f9-4589-bf12-a52425eaf9ad {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.149356] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1085.149356] env[62000]: value = "task-882925" [ 1085.149356] env[62000]: _type = "Task" [ 1085.149356] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.161713] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882925, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.292159] env[62000]: DEBUG nova.network.neutron [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.661026] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882925, 'name': PowerOffVM_Task, 'duration_secs': 0.230681} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.661465] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1085.662549] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6228eb5-6beb-4ef8-a848-be6da7891404 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.686089] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eba62c6d-6131-4571-baec-23c410c5ed62 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.795916] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.796359] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updated the network info_cache for instance {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9997}} [ 1085.797210] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.798886] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1085.798886] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1085.798886] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1086.175690] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1086.176699] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6888639a-e990-463e-a92a-09908fc9807a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.182872] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1086.183075] env[62000]: ERROR oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk due to incomplete transfer. [ 1086.183330] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-bdd246b3-9a57-4d8d-8b35-f65554aeb6a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.190899] env[62000]: DEBUG oslo_vmware.rw_handles [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5298a8ac-7d28-7524-7088-2a83c6998477/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1086.191129] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Uploaded image 38a40578-3968-46a5-b445-a5b08f645b60 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1086.193526] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1086.193786] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-76443d47-16ad-4237-bf4b-52c4da863528 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.198419] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Creating Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1086.198675] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-8206e8b7-b39c-4252-ba2d-894114ce051f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.201903] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1086.201903] env[62000]: value = "task-882926" [ 1086.201903] env[62000]: _type = "Task" [ 1086.201903] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.206138] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1086.206138] env[62000]: value = "task-882927" [ 1086.206138] env[62000]: _type = "Task" [ 1086.206138] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.212630] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882926, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.220085] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882927, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.301899] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.302041] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.302344] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.302523] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1086.303666] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94f05e70-1dc1-4e8d-831f-bb227f79e045 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.311252] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6e685a-c446-4e07-b20a-57e0eaa907fe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.325955] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781c8542-9289-4e9a-9940-e4416bfa2c0a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.333016] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f63d6d5f-39e3-4c9c-ac88-7b547eeed42d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.362443] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180617MB free_disk=52GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1086.362619] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.362796] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.716536] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882926, 'name': Destroy_Task, 'duration_secs': 0.312504} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.717197] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Destroyed the VM [ 1086.717452] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1086.717701] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-a1575896-706f-436c-ad6f-efb2300fdaf9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.721977] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882927, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.729385] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1086.729385] env[62000]: value = "task-882928" [ 1086.729385] env[62000]: _type = "Task" [ 1086.729385] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.738269] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882928, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.217927] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882927, 'name': CreateSnapshot_Task, 'duration_secs': 0.884957} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.218237] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Created Snapshot of the VM instance {{(pid=62000) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1087.219137] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d7c735-7e39-497e-aa1c-dad25d2b83c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.236820] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882928, 'name': RemoveSnapshot_Task, 'duration_secs': 0.366242} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.237140] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1087.237674] env[62000]: DEBUG nova.compute.manager [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1087.238139] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0ec65d-8dbc-4d1b-a3a3-df5c542c03c4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.514784] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 15fde663-fac4-4198-962f-8f814f1317f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1087.514784] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 19410eb7-0fec-4270-89da-04a2975fc050 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1087.514914] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 58b4c1b2-11b2-4d52-91c6-6df943b500f3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1087.515174] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1087.515228] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1087.569890] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3011c283-6593-4997-9bbe-7b429459c46e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.577293] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2effbd1c-e105-46e8-b343-3a5faed56e58 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.608595] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97576d5f-458f-4258-8dd0-4f6c9f57710e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.614799] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72b7db30-3561-4e88-8f23-ee376a0c7d30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.628664] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.736583] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Creating linked-clone VM from snapshot {{(pid=62000) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1087.736923] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f2ffabfc-af8d-4cb1-a7d2-b20315566a0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.748816] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1087.748816] env[62000]: value = "task-882930" [ 1087.748816] env[62000]: _type = "Task" [ 1087.748816] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.749292] env[62000]: INFO nova.compute.manager [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Shelve offloading [ 1087.755044] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1087.755044] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96ba6819-ef02-44a3-b02b-1476945f09af {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.761503] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882930, 'name': CloneVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.762750] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1087.762750] env[62000]: value = "task-882931" [ 1087.762750] env[62000]: _type = "Task" [ 1087.762750] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.771286] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882931, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.132091] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.260719] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882930, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.272685] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1088.272943] env[62000]: DEBUG nova.compute.manager [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.273774] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e31bdae-3294-4002-853d-7c535999f61c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.279853] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.280051] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.280238] env[62000]: DEBUG nova.network.neutron [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1088.491368] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "c7d808d7-fa85-484c-822b-25f3cda5090b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.491651] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.516323] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.516686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.639619] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1088.639882] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.277s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.761915] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882930, 'name': CloneVM_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.872274] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1088.872517] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1088.997422] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.018763] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1089.030188] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1089.057902] env[62000]: DEBUG nova.network.neutron [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1089.261478] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882930, 'name': CloneVM_Task, 'duration_secs': 1.396009} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.261668] env[62000]: INFO nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Created linked-clone VM from snapshot [ 1089.262463] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f969f694-a141-481c-a8e5-37cefb19fdab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.271023] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Uploading image 74362b29-2b5c-4427-a86f-be14376f4ef3 {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:442}} [ 1089.300421] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1089.300421] env[62000]: value = "vm-201655" [ 1089.300421] env[62000]: _type = "VirtualMachine" [ 1089.300421] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1089.300726] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-24ba6ec2-ac29-4fe6-9de6-d85c36531d74 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.307355] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease: (returnval){ [ 1089.307355] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205e951-7510-9aff-014e-302c7ae552d3" [ 1089.307355] env[62000]: _type = "HttpNfcLease" [ 1089.307355] env[62000]: } obtained for exporting VM: (result){ [ 1089.307355] env[62000]: value = "vm-201655" [ 1089.307355] env[62000]: _type = "VirtualMachine" [ 1089.307355] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1089.307589] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the lease: (returnval){ [ 1089.307589] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205e951-7510-9aff-014e-302c7ae552d3" [ 1089.307589] env[62000]: _type = "HttpNfcLease" [ 1089.307589] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1089.313416] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1089.313416] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205e951-7510-9aff-014e-302c7ae552d3" [ 1089.313416] env[62000]: _type = "HttpNfcLease" [ 1089.313416] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1089.521154] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.521439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.523692] env[62000]: INFO nova.compute.claims [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1089.543008] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.558275] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.780176] env[62000]: DEBUG nova.compute.manager [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-vif-unplugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1089.780176] env[62000]: DEBUG oslo_concurrency.lockutils [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1089.780176] env[62000]: DEBUG oslo_concurrency.lockutils [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1089.780534] env[62000]: DEBUG oslo_concurrency.lockutils [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.780534] env[62000]: DEBUG nova.compute.manager [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] No waiting events found dispatching network-vif-unplugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1089.780623] env[62000]: WARNING nova.compute.manager [req-2f6f604b-59f9-44e4-be3f-7c2a01645a58 req-7d960b07-e6dc-4de8-8c12-718f04ccf555 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received unexpected event network-vif-unplugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 for instance with vm_state shelved and task_state shelving_offloading. [ 1089.815633] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1089.815633] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205e951-7510-9aff-014e-302c7ae552d3" [ 1089.815633] env[62000]: _type = "HttpNfcLease" [ 1089.815633] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1089.815927] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1089.815927] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5205e951-7510-9aff-014e-302c7ae552d3" [ 1089.815927] env[62000]: _type = "HttpNfcLease" [ 1089.815927] env[62000]: }. {{(pid=62000) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1089.816707] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7abe310c-edf9-4be4-9f92-34cd8240339c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.823765] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1089.823943] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk for reading. {{(pid=62000) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1089.883243] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1089.884163] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39546199-f6aa-4e84-891a-bb8c95f09101 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.893315] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1089.893618] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ba82587-1018-4279-be45-9e51616ca8b9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.937155] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-58839592-8840-40d5-8ebd-434d76349702 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.963937] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1089.964315] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1089.964518] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1089.969516] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-32f6cf61-74bc-482e-a328-44a1557a7c0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.975762] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1089.975762] env[62000]: value = "task-882934" [ 1089.975762] env[62000]: _type = "Task" [ 1089.975762] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.984400] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882934, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.486058] env[62000]: DEBUG oslo_vmware.api [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882934, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147161} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.486765] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1090.487104] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1090.487349] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1090.515685] env[62000]: INFO nova.scheduler.client.report [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted allocations for instance 15fde663-fac4-4198-962f-8f814f1317f6 [ 1090.613787] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9e8dc8-8fbe-4b68-a52b-7d87c1b3fef9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.622630] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc563288-9e33-4e45-bc63-da907f87e0fd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.655852] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea248da6-2a73-4701-a665-d2d67ffadc4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.663386] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed27c49d-441e-4d3d-9d52-166986ec4a6e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.677103] env[62000]: DEBUG nova.compute.provider_tree [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1091.023747] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1091.033259] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1091.033674] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Cleaning up deleted instances {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11226}} [ 1091.181769] env[62000]: DEBUG nova.scheduler.client.report [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1091.556998] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] There are 60 instances to clean {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11235}} [ 1091.557408] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 6d2eaac3-c707-4850-8fc7-9edd56c8bc17] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1091.689125] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.168s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1091.689826] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1091.692857] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.150s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1091.694835] env[62000]: INFO nova.compute.claims [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1091.805941] env[62000]: DEBUG nova.compute.manager [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1091.806242] env[62000]: DEBUG nova.compute.manager [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing instance network info cache due to event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1091.806622] env[62000]: DEBUG oslo_concurrency.lockutils [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.806801] env[62000]: DEBUG oslo_concurrency.lockutils [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.807179] env[62000]: DEBUG nova.network.neutron [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1092.061880] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 67af85f1-9134-440d-a5df-09ec7d3e72a0] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.199163] env[62000]: DEBUG nova.compute.utils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1092.204607] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1092.204925] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1092.257081] env[62000]: DEBUG nova.policy [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b04f86982584d59b613b508f99d2c12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15b98199f74fdeb39abb58210f1c0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1092.564376] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 9ee49e0b-6d37-4826-8d8a-bfb3752af4f8] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1092.707685] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1092.739420] env[62000]: DEBUG nova.network.neutron [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updated VIF entry in instance network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1092.739420] env[62000]: DEBUG nova.network.neutron [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.765763] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Successfully created port: cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1092.801185] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7c3f7f2-4346-489e-a9b7-131dc6c340b6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.809589] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282d7a51-c47f-405a-90b2-cb6e56572c4c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.840105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b8e7e0-8b08-4d38-a937-7040896c573c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.847612] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b3cad5-3097-4a65-8bfd-0a1ba0c79d4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.861166] env[62000]: DEBUG nova.compute.provider_tree [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.071645] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 67f8274b-a0ad-419a-81fc-515b06ad41aa] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.241368] env[62000]: DEBUG oslo_concurrency.lockutils [req-f4710596-fd74-4dd2-8811-1c708bbdd3ef req-cf2bfc01-684e-4ffb-a3f6-5469f88f2a50 service nova] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.364058] env[62000]: DEBUG nova.scheduler.client.report [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.575017] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 64ca26bd-dc8c-4f00-bfde-a24f8d650848] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1093.718261] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1093.746106] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1093.746492] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1093.746674] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1093.746869] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1093.747125] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1093.747343] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1093.747583] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1093.747768] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1093.748057] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1093.748268] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1093.748454] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1093.749469] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb1c28d-872c-4aae-9807-168c3e8bd7f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.758431] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39777ff4-8351-4670-ae61-23d5a23978e9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.793588] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.869953] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.177s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1093.870462] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1093.873485] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.850s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.873747] env[62000]: DEBUG nova.objects.instance [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'resources' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.078271] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 9ccf07fa-90db-40b9-af5d-b47ee5773f9c] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.162836] env[62000]: DEBUG nova.compute.manager [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Received event network-vif-plugged-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1094.163095] env[62000]: DEBUG oslo_concurrency.lockutils [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] Acquiring lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1094.163338] env[62000]: DEBUG oslo_concurrency.lockutils [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.163543] env[62000]: DEBUG oslo_concurrency.lockutils [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.163728] env[62000]: DEBUG nova.compute.manager [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] No waiting events found dispatching network-vif-plugged-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1094.163845] env[62000]: WARNING nova.compute.manager [req-011bb86b-57a8-48ee-82f3-72943032eec2 req-d8fa369c-a386-4240-9d41-8839656476f1 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Received unexpected event network-vif-plugged-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 for instance with vm_state building and task_state spawning. [ 1094.251674] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Successfully updated port: cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1094.376786] env[62000]: DEBUG nova.objects.instance [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'numa_topology' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.378897] env[62000]: DEBUG nova.compute.utils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1094.380406] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1094.380602] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1094.420948] env[62000]: DEBUG nova.policy [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b04f86982584d59b613b508f99d2c12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15b98199f74fdeb39abb58210f1c0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1094.582482] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: a2166191-09b7-4e5b-9cca-521f76814fb2] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1094.673751] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Successfully created port: 8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1094.754161] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.754311] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.754478] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1094.881012] env[62000]: DEBUG nova.objects.base [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Object Instance<15fde663-fac4-4198-962f-8f814f1317f6> lazy-loaded attributes: resources,numa_topology {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1094.884346] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1094.969690] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48ace845-ed44-4336-bc13-bccfbb2846a7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.977849] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45b04c6d-320d-4373-ad6b-b018c8fa6ca7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.008018] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b549b9a-f1c7-43a0-aba0-0286a2f44454 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.016059] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd650ef8-b50a-4861-a6e3-d5499a781030 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.029376] env[62000]: DEBUG nova.compute.provider_tree [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.085803] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 9f864cad-f454-4dc3-b6d0-793d7adec0ae] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.298444] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1095.430265] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Updating instance_info_cache with network_info: [{"id": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "address": "fa:16:3e:5b:41:18", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdfe8b13-6a", "ovs_interfaceid": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.532883] env[62000]: DEBUG nova.scheduler.client.report [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.589281] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 0246b032-505e-4bc5-bfc0-5779ff564626] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1095.894738] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1095.919757] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1095.920094] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1095.920300] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1095.920496] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1095.920647] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1095.920802] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1095.920997] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1095.921182] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1095.921387] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1095.921514] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1095.921680] env[62000]: DEBUG nova.virt.hardware [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1095.922618] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab028bd1-df8a-4c19-af73-7a9f22263eec {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.930492] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2986ea3-bef9-4cec-b9f3-9a089256e39c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.934559] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.934877] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Instance network_info: |[{"id": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "address": "fa:16:3e:5b:41:18", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdfe8b13-6a", "ovs_interfaceid": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1095.935276] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:5b:41:18', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089a7624-43ba-4fce-bfc0-63e4bb7f9aeb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1095.942579] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating folder: Project (aa15b98199f74fdeb39abb58210f1c0e). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1095.943250] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4ccbf182-0225-4eaf-9758-e134014617a5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.961364] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created folder: Project (aa15b98199f74fdeb39abb58210f1c0e) in parent group-v201431. [ 1095.961364] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating folder: Instances. Parent ref: group-v201656. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1095.961595] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fb97bba-1a03-4021-af58-fea086f6ae4f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.970488] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created folder: Instances in parent group-v201656. [ 1095.970731] env[62000]: DEBUG oslo.service.loopingcall [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1095.970930] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1095.971178] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d93de2c9-2948-41aa-b275-0d81f614e1ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.989226] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1095.989226] env[62000]: value = "task-882937" [ 1095.989226] env[62000]: _type = "Task" [ 1095.989226] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.996342] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882937, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.037579] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.164s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.092208] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f8f26be0-f7d1-43e9-886d-c19d385e6935] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.190868] env[62000]: DEBUG nova.compute.manager [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Received event network-changed-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1096.191121] env[62000]: DEBUG nova.compute.manager [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Refreshing instance network info cache due to event network-changed-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1096.191366] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Acquiring lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.191519] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Acquired lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.191709] env[62000]: DEBUG nova.network.neutron [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Refreshing network info cache for port cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1096.256737] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Successfully updated port: 8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1096.498707] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882937, 'name': CreateVM_Task, 'duration_secs': 0.422304} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.498947] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1096.499638] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.499825] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.500186] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1096.500452] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79da17ad-6c5a-49c0-8411-ecd078134eaf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.504987] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1096.504987] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994634-f396-997e-2ebb-1ebd8efefeb8" [ 1096.504987] env[62000]: _type = "Task" [ 1096.504987] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.512222] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994634-f396-997e-2ebb-1ebd8efefeb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.546302] env[62000]: DEBUG oslo_concurrency.lockutils [None req-1efd3794-a279-4b58-9e1c-dfa0ea772aa7 tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.289s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.547209] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.754s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.547462] env[62000]: INFO nova.compute.manager [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Unshelving [ 1096.595782] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: c9758d78-83a0-430e-a69e-e2d82f99daa9] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1096.760061] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1096.760061] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1096.760061] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1096.892013] env[62000]: DEBUG nova.network.neutron [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Updated VIF entry in instance network info cache for port cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1096.892496] env[62000]: DEBUG nova.network.neutron [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Updating instance_info_cache with network_info: [{"id": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "address": "fa:16:3e:5b:41:18", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdfe8b13-6a", "ovs_interfaceid": "cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.015446] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994634-f396-997e-2ebb-1ebd8efefeb8, 'name': SearchDatastore_Task, 'duration_secs': 0.030789} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.015759] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.015999] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1097.016255] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1097.016439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1097.016626] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1097.016899] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d1955262-a82d-4f51-a471-d15489a1eca0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.026440] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1097.026644] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1097.027576] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1f61ef1-01ad-455f-81ce-d82bc56365e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.033415] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1097.033415] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522e2e6c-b2e4-3e4d-a423-237689cd8b0f" [ 1097.033415] env[62000]: _type = "Task" [ 1097.033415] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.041110] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522e2e6c-b2e4-3e4d-a423-237689cd8b0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.098853] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 0e66a948-61d5-4991-99bd-374db7eaaf2a] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.300675] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1097.395573] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Releasing lock "refresh_cache-c7d808d7-fa85-484c-822b-25f3cda5090b" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.395887] env[62000]: DEBUG nova.compute.manager [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Received event network-vif-plugged-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1097.396134] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Acquiring lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.396371] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.396565] env[62000]: DEBUG oslo_concurrency.lockutils [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1097.396744] env[62000]: DEBUG nova.compute.manager [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] No waiting events found dispatching network-vif-plugged-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1097.396947] env[62000]: WARNING nova.compute.manager [req-cfcc00e4-32b1-44bf-8db6-ef091c09483a req-2bdb9d37-16f8-4fcf-8c0d-b90f77846156 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Received unexpected event network-vif-plugged-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad for instance with vm_state building and task_state spawning. [ 1097.441079] env[62000]: DEBUG nova.network.neutron [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Updating instance_info_cache with network_info: [{"id": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "address": "fa:16:3e:c5:48:16", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab50541-4f", "ovs_interfaceid": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1097.544175] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]522e2e6c-b2e4-3e4d-a423-237689cd8b0f, 'name': SearchDatastore_Task, 'duration_secs': 0.011178} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.545050] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5cc9092-7bfc-4ada-aef0-219bceb27fe9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.550347] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1097.550347] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52046e1e-57f7-6adc-d167-b1dec0cb6058" [ 1097.550347] env[62000]: _type = "Task" [ 1097.550347] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.561081] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52046e1e-57f7-6adc-d167-b1dec0cb6058, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.574385] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1097.574685] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1097.574920] env[62000]: DEBUG nova.objects.instance [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'pci_requests' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.603030] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 23366d62-294b-47c8-adc6-80e854f452ef] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1097.944015] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1097.944456] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Instance network_info: |[{"id": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "address": "fa:16:3e:c5:48:16", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab50541-4f", "ovs_interfaceid": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1097.944940] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:48:16', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089a7624-43ba-4fce-bfc0-63e4bb7f9aeb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1097.952790] env[62000]: DEBUG oslo.service.loopingcall [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1097.953062] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1097.953344] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31fd460d-7cec-4429-a3b7-0874e6761347 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.974146] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1097.974146] env[62000]: value = "task-882938" [ 1097.974146] env[62000]: _type = "Task" [ 1097.974146] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.981746] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882938, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.060396] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52046e1e-57f7-6adc-d167-b1dec0cb6058, 'name': SearchDatastore_Task, 'duration_secs': 0.011606} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.060669] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1098.060939] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c7d808d7-fa85-484c-822b-25f3cda5090b/c7d808d7-fa85-484c-822b-25f3cda5090b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1098.061220] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb2e1cbb-9bf2-4849-9682-6587273dfd6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.067477] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1098.067477] env[62000]: value = "task-882939" [ 1098.067477] env[62000]: _type = "Task" [ 1098.067477] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.074877] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.078596] env[62000]: DEBUG nova.objects.instance [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'numa_topology' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1098.106827] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: b6579b75-42dc-473d-8b66-24aa39f9e4ff] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.166974] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1098.168112] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372ae06c-4a34-49a5-a479-d119177cccfc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.175406] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1098.175674] env[62000]: ERROR oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk due to incomplete transfer. [ 1098.175878] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-b575c2e3-5580-429b-bbd7-3310a5dca14c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.183186] env[62000]: DEBUG oslo_vmware.rw_handles [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52ef2729-d0f3-c36d-2755-0b95843637cb/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1098.183442] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Uploaded image 74362b29-2b5c-4427-a86f-be14376f4ef3 to the Glance image server {{(pid=62000) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:473}} [ 1098.186387] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Destroying the VM {{(pid=62000) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1358}} [ 1098.186701] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-10c248b6-4a28-4545-b584-ca51cf4c2f16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.193152] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1098.193152] env[62000]: value = "task-882940" [ 1098.193152] env[62000]: _type = "Task" [ 1098.193152] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.201869] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882940, 'name': Destroy_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.218090] env[62000]: DEBUG nova.compute.manager [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Received event network-changed-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1098.218312] env[62000]: DEBUG nova.compute.manager [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Refreshing instance network info cache due to event network-changed-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1098.218549] env[62000]: DEBUG oslo_concurrency.lockutils [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] Acquiring lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.218672] env[62000]: DEBUG oslo_concurrency.lockutils [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] Acquired lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.218838] env[62000]: DEBUG nova.network.neutron [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Refreshing network info cache for port 8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1098.484570] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882938, 'name': CreateVM_Task, 'duration_secs': 0.418426} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.484745] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1098.485463] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1098.485642] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1098.485986] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1098.486281] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13b29e27-3250-4ada-9395-12e18c947e45 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.491429] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1098.491429] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5292ad91-522c-c1a1-14aa-1250ca055a5e" [ 1098.491429] env[62000]: _type = "Task" [ 1098.491429] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.499596] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5292ad91-522c-c1a1-14aa-1250ca055a5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.576352] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882939, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.580869] env[62000]: INFO nova.compute.claims [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1098.610338] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: ea58da8c-e4ac-4863-942d-6294cf04e6f2] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1098.703459] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882940, 'name': Destroy_Task} progress is 33%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.914744] env[62000]: DEBUG nova.network.neutron [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Updated VIF entry in instance network info cache for port 8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1098.915154] env[62000]: DEBUG nova.network.neutron [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Updating instance_info_cache with network_info: [{"id": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "address": "fa:16:3e:c5:48:16", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8ab50541-4f", "ovs_interfaceid": "8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.001359] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5292ad91-522c-c1a1-14aa-1250ca055a5e, 'name': SearchDatastore_Task, 'duration_secs': 0.04612} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.001652] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.001889] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1099.002174] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1099.002333] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1099.002519] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1099.002774] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ae428f0-3a2f-4b94-bff1-283aee233d8a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.010032] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1099.010263] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1099.010914] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-125e277d-e44e-4221-8f51-93225df9abab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.015373] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1099.015373] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5221ce96-5a95-baab-2504-55a873b874dc" [ 1099.015373] env[62000]: _type = "Task" [ 1099.015373] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.022144] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5221ce96-5a95-baab-2504-55a873b874dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.076463] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526412} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.076704] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] c7d808d7-fa85-484c-822b-25f3cda5090b/c7d808d7-fa85-484c-822b-25f3cda5090b.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1099.076946] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1099.077203] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dff6ad11-d00f-46ec-bfa7-050401e64444 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.083543] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1099.083543] env[62000]: value = "task-882941" [ 1099.083543] env[62000]: _type = "Task" [ 1099.083543] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.093695] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882941, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.113963] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 54eb64ad-bc7b-4495-9e93-451059139c4c] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.203057] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882940, 'name': Destroy_Task, 'duration_secs': 0.570706} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.203535] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Destroyed the VM [ 1099.203786] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleting Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1099.204046] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-759b2cfa-3b9b-4965-ad2d-4b0ec0ddad26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.209583] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1099.209583] env[62000]: value = "task-882942" [ 1099.209583] env[62000]: _type = "Task" [ 1099.209583] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.217130] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882942, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.417849] env[62000]: DEBUG oslo_concurrency.lockutils [req-4678a1d9-2f46-4cee-945b-c44fa9eadb24 req-1d45c65e-d7fb-4259-a683-c494cdcdf71b service nova] Releasing lock "refresh_cache-e0c78a4a-fe40-46cd-b0f3-dac971866cba" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1099.526373] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5221ce96-5a95-baab-2504-55a873b874dc, 'name': SearchDatastore_Task, 'duration_secs': 0.007112} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.527183] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4208f6cf-d4ba-45d1-86fd-6f60140ab99e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.532815] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1099.532815] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52863a28-eb7c-80f7-a764-1334bf845b0d" [ 1099.532815] env[62000]: _type = "Task" [ 1099.532815] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.540746] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52863a28-eb7c-80f7-a764-1334bf845b0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.594405] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882941, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064615} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.594811] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1099.595575] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00638a0-4cdc-4cd3-93e3-b0a91b900dac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.616801] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] c7d808d7-fa85-484c-822b-25f3cda5090b/c7d808d7-fa85-484c-822b-25f3cda5090b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1099.619324] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 3878579f-6435-4fe3-9f8c-8461d8ac57ee] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1099.620988] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fb7c5f8-c68d-4b0a-b672-f106e7b8bd95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.640268] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1099.640268] env[62000]: value = "task-882943" [ 1099.640268] env[62000]: _type = "Task" [ 1099.640268] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.647944] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882943, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.690162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-613f3289-ebe8-4a7d-8e7a-c3482d1de138 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.697418] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-890b72cb-8c69-4af1-b057-28e876954149 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.730553] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd3824b-d980-475e-a57f-2ff85d4cea93 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.737276] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882942, 'name': RemoveSnapshot_Task, 'duration_secs': 0.367281} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.739183] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleted Snapshot of the VM instance {{(pid=62000) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1099.739486] env[62000]: DEBUG nova.compute.manager [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1099.740216] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45f71d5d-a7ff-4333-9ff1-07474e117132 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.743287] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29387a32-e9b0-46d9-8423-c43d4e7e63ea {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.759047] env[62000]: DEBUG nova.compute.provider_tree [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.042639] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52863a28-eb7c-80f7-a764-1334bf845b0d, 'name': SearchDatastore_Task, 'duration_secs': 0.019266} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.043262] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1100.043527] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] e0c78a4a-fe40-46cd-b0f3-dac971866cba/e0c78a4a-fe40-46cd-b0f3-dac971866cba.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1100.043792] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02df0ac4-1ec1-489e-a4b6-6c34677a2f6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.051475] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1100.051475] env[62000]: value = "task-882944" [ 1100.051475] env[62000]: _type = "Task" [ 1100.051475] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.058911] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882944, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.136691] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: bf606000-346b-48db-972a-000b54a8ec5d] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.149610] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882943, 'name': ReconfigVM_Task, 'duration_secs': 0.242346} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.149908] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Reconfigured VM instance instance-0000006b to attach disk [datastore1] c7d808d7-fa85-484c-822b-25f3cda5090b/c7d808d7-fa85-484c-822b-25f3cda5090b.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1100.150651] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-693e44ea-8b48-42a9-a1d5-fbef701c8e56 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.156800] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1100.156800] env[62000]: value = "task-882945" [ 1100.156800] env[62000]: _type = "Task" [ 1100.156800] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.165485] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882945, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.264109] env[62000]: DEBUG nova.scheduler.client.report [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1100.267715] env[62000]: INFO nova.compute.manager [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Shelve offloading [ 1100.269560] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1100.269810] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c37367c6-c1fc-411f-bc19-075cb9c29169 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.276471] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1100.276471] env[62000]: value = "task-882946" [ 1100.276471] env[62000]: _type = "Task" [ 1100.276471] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.286328] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] VM already powered off {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1509}} [ 1100.286562] env[62000]: DEBUG nova.compute.manager [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1100.287404] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4819b024-d7b0-4bb7-a267-ac622fa8d617 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.293945] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1100.294141] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1100.294328] env[62000]: DEBUG nova.network.neutron [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1100.561550] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882944, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.428993} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.561826] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] e0c78a4a-fe40-46cd-b0f3-dac971866cba/e0c78a4a-fe40-46cd-b0f3-dac971866cba.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1100.562113] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1100.562372] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dbb31e72-2ced-40f2-97ea-a408fa673b30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.569914] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1100.569914] env[62000]: value = "task-882947" [ 1100.569914] env[62000]: _type = "Task" [ 1100.569914] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.577286] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882947, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.640092] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: dfa26fab-40ff-4409-bb73-79aa555ab225] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1100.667053] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882945, 'name': Rename_Task, 'duration_secs': 0.368542} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.667329] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1100.667644] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cf6d82c4-8291-443f-b392-4aab42cd4730 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.673187] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1100.673187] env[62000]: value = "task-882948" [ 1100.673187] env[62000]: _type = "Task" [ 1100.673187] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1100.680243] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882948, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1100.770129] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.195s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1100.799293] env[62000]: INFO nova.network.neutron [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1101.080536] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882947, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06185} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.083093] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1101.083625] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be45168-4786-4906-9856-4c771c28637c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.105398] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] e0c78a4a-fe40-46cd-b0f3-dac971866cba/e0c78a4a-fe40-46cd-b0f3-dac971866cba.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1101.105654] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1eb89180-ad34-4caf-9a4f-804a821a06a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.124631] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1101.124631] env[62000]: value = "task-882949" [ 1101.124631] env[62000]: _type = "Task" [ 1101.124631] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.132204] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882949, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.141901] env[62000]: DEBUG nova.network.neutron [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1101.143269] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: a0e2344f-32ec-4fb1-ba7c-99e18f422923] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.182890] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882948, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.635507] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882949, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.646105] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1101.648006] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 94605ab8-8167-430e-b1cd-c8f51e50d8b9] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1101.684110] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882948, 'name': PowerOnVM_Task, 'duration_secs': 0.664697} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.684396] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1101.684597] env[62000]: INFO nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Took 7.97 seconds to spawn the instance on the hypervisor. [ 1101.684778] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1101.685545] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ac3caee-ca52-4eb5-8a73-e360ea56a489 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.858214] env[62000]: DEBUG nova.compute.manager [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-vif-unplugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1101.858494] env[62000]: DEBUG oslo_concurrency.lockutils [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.858683] env[62000]: DEBUG oslo_concurrency.lockutils [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.858822] env[62000]: DEBUG oslo_concurrency.lockutils [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.858996] env[62000]: DEBUG nova.compute.manager [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] No waiting events found dispatching network-vif-unplugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1101.859271] env[62000]: WARNING nova.compute.manager [req-ab14edf2-063a-47d0-8fb8-6f9584a925dd req-f850b469-e4ca-4720-aea1-a103eaae7a7d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received unexpected event network-vif-unplugged-69c33d12-7f23-4311-89f4-5b66d2fc837e for instance with vm_state shelved and task_state shelving_offloading. [ 1101.945501] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1101.946596] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18161e97-5ac7-451c-8c60-292d37348fb9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.954175] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1101.954408] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f9b88979-603e-41f7-8086-64ed70798191 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.048856] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1102.049114] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1102.049322] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleting the datastore file [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1102.049627] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a3a3c94-1f08-4d73-a576-4b532afe4287 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.057715] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1102.057715] env[62000]: value = "task-882951" [ 1102.057715] env[62000]: _type = "Task" [ 1102.057715] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.065063] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882951, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.135718] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882949, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.151358] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 2a81afa3-bca1-4ccc-a1a8-3cd9dbf94283] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.202170] env[62000]: INFO nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Took 12.70 seconds to build instance. [ 1102.566934] env[62000]: DEBUG oslo_vmware.api [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882951, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.301691} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.567295] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.567553] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1102.567796] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1102.591203] env[62000]: INFO nova.scheduler.client.report [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted allocations for instance 19410eb7-0fec-4270-89da-04a2975fc050 [ 1102.636728] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882949, 'name': ReconfigVM_Task, 'duration_secs': 1.232579} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.637060] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Reconfigured VM instance instance-0000006c to attach disk [datastore1] e0c78a4a-fe40-46cd-b0f3-dac971866cba/e0c78a4a-fe40-46cd-b0f3-dac971866cba.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1102.637865] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-49a2d0ee-bb73-4a3a-b6fe-638177e3a9d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.644119] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1102.644119] env[62000]: value = "task-882952" [ 1102.644119] env[62000]: _type = "Task" [ 1102.644119] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1102.646189] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.646189] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.646189] env[62000]: DEBUG nova.network.neutron [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1102.656526] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 0f5b416a-20e4-42f5-9ad9-a8489ab11e3b] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1102.658803] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882952, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.704217] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.212s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.097650] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.097924] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.098193] env[62000]: DEBUG nova.objects.instance [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'resources' on Instance uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.159472] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: c354ba41-0915-44cc-9857-42705a56c00f] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.161281] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882952, 'name': Rename_Task, 'duration_secs': 0.13135} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.161539] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1103.161816] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5838ebd5-e63b-4a0d-95fb-ad461601ad47 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.167567] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1103.167567] env[62000]: value = "task-882953" [ 1103.167567] env[62000]: _type = "Task" [ 1103.167567] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.176218] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882953, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.383096] env[62000]: DEBUG nova.network.neutron [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.601013] env[62000]: DEBUG nova.objects.instance [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'numa_topology' on Instance uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1103.665583] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: d2f87d82-af5e-4b17-959e-40865a8852b3] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1103.678289] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882953, 'name': PowerOnVM_Task} progress is 87%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.886755] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.890257] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1103.890466] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing instance network info cache due to event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1103.890678] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1103.890848] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1103.891009] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1103.919373] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='3268a490aa6d95f645975b871315c9c8',container_format='bare',created_at=2024-09-16T07:02:10Z,direct_url=,disk_format='vmdk',id=38a40578-3968-46a5-b445-a5b08f645b60,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-1081512547-shelved',owner='cac5f0a5704d434082131155e107d190',properties=ImageMetaProps,protected=,size=31668224,status='active',tags=,updated_at=2024-09-16T07:02:25Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1103.919640] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1103.919804] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1103.920048] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1103.920218] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1103.920377] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1103.920594] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1103.920761] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1103.920933] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1103.921146] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1103.921361] env[62000]: DEBUG nova.virt.hardware [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1103.922303] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17b788e1-bf32-4848-9499-84db31c3796e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.930909] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d11a19-7865-49ba-b951-74d4ec11e14e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.944948] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1e:1b:2b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9b7e9e55-3210-4fae-9648-d87e76c3d931', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2ec45f71-6b9c-4a0a-bd0f-f73708c92053', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1103.952557] env[62000]: DEBUG oslo.service.loopingcall [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1103.952862] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1103.953206] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3da12288-1d7d-4d69-ac7b-9c163e978ebf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.973716] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1103.973716] env[62000]: value = "task-882954" [ 1103.973716] env[62000]: _type = "Task" [ 1103.973716] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.981514] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882954, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.099488] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.103187] env[62000]: DEBUG nova.objects.base [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Object Instance<19410eb7-0fec-4270-89da-04a2975fc050> lazy-loaded attributes: resources,numa_topology {{(pid=62000) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1104.168754] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 72a5bf70-dc6e-4887-abb8-8fbad64bb065] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.182423] env[62000]: DEBUG oslo_vmware.api [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882953, 'name': PowerOnVM_Task, 'duration_secs': 0.657037} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.184444] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1104.184855] env[62000]: INFO nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Took 8.29 seconds to spawn the instance on the hypervisor. [ 1104.185207] env[62000]: DEBUG nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1104.186449] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-129b2b3e-35b3-4d1f-9a07-4b26c20dc906 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.192176] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b1b7a7f-017a-466e-b053-129b28c78041 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.206084] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9280abba-40e5-44d6-946d-e3a61afb15e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.263401] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff7802d8-4b63-48ef-a8ba-aff23924194c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.272013] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b00e668-ed3d-4566-8e48-6fa4b3d54db3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.288384] env[62000]: DEBUG nova.compute.provider_tree [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.486365] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882954, 'name': CreateVM_Task} progress is 25%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.675195] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 9097dd44-563c-460b-b718-cbc1d5eb4a28] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1104.703489] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updated VIF entry in instance network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1104.703860] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap69c33d12-7f", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.723490] env[62000]: INFO nova.compute.manager [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Took 15.20 seconds to build instance. [ 1104.791565] env[62000]: DEBUG nova.scheduler.client.report [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.985912] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882954, 'name': CreateVM_Task, 'duration_secs': 0.689898} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.986171] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1104.987014] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.987209] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.987601] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1104.987863] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-60265db5-4a49-48ef-8da7-29ce773927f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.992463] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1104.992463] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521d3989-842e-0933-da62-572933ebb4ec" [ 1104.992463] env[62000]: _type = "Task" [ 1104.992463] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.000059] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521d3989-842e-0933-da62-572933ebb4ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.178783] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: edc7c99e-d53c-4e0b-991d-c4b693544cdd] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.206253] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.206548] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.206765] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.206969] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.207158] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.207381] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] No waiting events found dispatching network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1105.207517] env[62000]: WARNING nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received unexpected event network-vif-plugged-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 for instance with vm_state shelved_offloaded and task_state spawning. [ 1105.207695] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1105.207912] env[62000]: DEBUG nova.compute.manager [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing instance network info cache due to event network-changed-2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1105.208090] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Acquiring lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.208251] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Acquired lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.208414] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Refreshing network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1105.225899] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af7fc514-4df8-414e-a0b0-c4636ab2ba07 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.709s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.296862] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.199s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.319921] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "c7d808d7-fa85-484c-822b-25f3cda5090b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.321386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.321386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.321386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.321386] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.323353] env[62000]: INFO nova.compute.manager [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Terminating instance [ 1105.325287] env[62000]: DEBUG nova.compute.manager [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1105.325490] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1105.326319] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff71a93d-d207-46f0-9435-85cad498ca3c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.334320] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1105.334572] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-606d5752-0e0e-4275-b063-0beb203d5fa1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.341473] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1105.341473] env[62000]: value = "task-882955" [ 1105.341473] env[62000]: _type = "Task" [ 1105.341473] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.349774] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882955, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.401604] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.401924] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.402350] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1105.402614] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.402815] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.405250] env[62000]: INFO nova.compute.manager [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Terminating instance [ 1105.407238] env[62000]: DEBUG nova.compute.manager [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1105.407458] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1105.408299] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1482a83-5ba2-4c09-af31-ea34cc1d4209 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.417441] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1105.417758] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08f7328f-c0c6-4fc2-8128-aa45de3afc30 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.424029] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1105.424029] env[62000]: value = "task-882956" [ 1105.424029] env[62000]: _type = "Task" [ 1105.424029] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.432581] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882956, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.504569] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.504947] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Processing image 38a40578-3968-46a5-b445-a5b08f645b60 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1105.505285] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1105.505492] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquired lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1105.505753] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1105.506073] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-73635847-ff3c-4380-8d0a-86a88a615477 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.513726] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1105.513974] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1105.514927] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-07550d93-1593-454c-b102-8cbd9bc3ebb8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.520802] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1105.520802] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c93569-2f47-e4a9-a892-68969217e586" [ 1105.520802] env[62000]: _type = "Task" [ 1105.520802] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.529524] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c93569-2f47-e4a9-a892-68969217e586, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.682399] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: cc7e8af9-16ef-41bf-a91e-edb6c166a3a5] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1105.805517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-3790ee1f-44ea-4239-9b9f-ccfb49781e42 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.175s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.805911] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.706s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.805911] env[62000]: INFO nova.compute.manager [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Unshelving [ 1105.852320] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882955, 'name': PowerOffVM_Task, 'duration_secs': 0.215725} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.852659] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1105.852851] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1105.853137] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-438ea7a3-d171-4db9-9b8f-e483f16db2eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.912177] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1105.912579] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1105.912671] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleting the datastore file [datastore1] c7d808d7-fa85-484c-822b-25f3cda5090b {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.912911] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de42a457-17b9-4022-b9e1-be985879b79a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.920052] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1105.920052] env[62000]: value = "task-882958" [ 1105.920052] env[62000]: _type = "Task" [ 1105.920052] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.930767] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882958, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.931571] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updated VIF entry in instance network info cache for port 2ec45f71-6b9c-4a0a-bd0f-f73708c92053. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1105.931947] env[62000]: DEBUG nova.network.neutron [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [{"id": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "address": "fa:16:3e:1e:1b:2b", "network": {"id": "62b85df3-226a-4b5a-bd60-4c3d262b3446", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-2034025614-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.195", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "cac5f0a5704d434082131155e107d190", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9b7e9e55-3210-4fae-9648-d87e76c3d931", "external-id": "nsx-vlan-transportzone-967", "segmentation_id": 967, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2ec45f71-6b", "ovs_interfaceid": "2ec45f71-6b9c-4a0a-bd0f-f73708c92053", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.937463] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882956, 'name': PowerOffVM_Task, 'duration_secs': 0.167774} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.937673] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1105.937849] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1105.938117] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37a8affc-65b4-443c-8157-907d86b8d5d8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.996671] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1105.996921] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1105.997148] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleting the datastore file [datastore1] e0c78a4a-fe40-46cd-b0f3-dac971866cba {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.997439] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c47f6a3d-e994-4cec-9063-882d9b6d0f85 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.004602] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1106.004602] env[62000]: value = "task-882960" [ 1106.004602] env[62000]: _type = "Task" [ 1106.004602] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.012763] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882960, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.032036] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1106.032357] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Fetch image to [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77/OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1106.032595] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Downloading stream optimized image 38a40578-3968-46a5-b445-a5b08f645b60 to [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77/OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77.vmdk on the data store datastore2 as vApp {{(pid=62000) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1106.032819] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Downloading image file data 38a40578-3968-46a5-b445-a5b08f645b60 to the ESX as VM named 'OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77' {{(pid=62000) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1106.118234] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1106.118234] env[62000]: value = "resgroup-9" [ 1106.118234] env[62000]: _type = "ResourcePool" [ 1106.118234] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1106.118618] env[62000]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-b5c0abdd-0c84-4a7b-9c86-86648ebb0af0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.140133] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease: (returnval){ [ 1106.140133] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b609a7-6249-d3c8-3cee-7ee8244cefa9" [ 1106.140133] env[62000]: _type = "HttpNfcLease" [ 1106.140133] env[62000]: } obtained for vApp import into resource pool (val){ [ 1106.140133] env[62000]: value = "resgroup-9" [ 1106.140133] env[62000]: _type = "ResourcePool" [ 1106.140133] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1106.140578] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the lease: (returnval){ [ 1106.140578] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b609a7-6249-d3c8-3cee-7ee8244cefa9" [ 1106.140578] env[62000]: _type = "HttpNfcLease" [ 1106.140578] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1106.146137] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1106.146137] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b609a7-6249-d3c8-3cee-7ee8244cefa9" [ 1106.146137] env[62000]: _type = "HttpNfcLease" [ 1106.146137] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1106.186080] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 44a54bd7-a35c-49ed-85ed-346830cee6ad] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.430512] env[62000]: DEBUG oslo_vmware.api [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882958, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.183028} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.430774] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.431028] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1106.431236] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1106.431348] env[62000]: INFO nova.compute.manager [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1106.431596] env[62000]: DEBUG oslo.service.loopingcall [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.431789] env[62000]: DEBUG nova.compute.manager [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1106.431882] env[62000]: DEBUG nova.network.neutron [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1106.434083] env[62000]: DEBUG oslo_concurrency.lockutils [req-19cdcb3d-024b-415b-88b0-0c68b1d4ce43 req-d2c1f802-af62-4765-9f97-84994c2ae80e service nova] Releasing lock "refresh_cache-15fde663-fac4-4198-962f-8f814f1317f6" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1106.514253] env[62000]: DEBUG oslo_vmware.api [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882960, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.19633} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.514583] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1106.514777] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1106.514958] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1106.515153] env[62000]: INFO nova.compute.manager [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1106.515393] env[62000]: DEBUG oslo.service.loopingcall [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1106.515582] env[62000]: DEBUG nova.compute.manager [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1106.515677] env[62000]: DEBUG nova.network.neutron [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1106.651194] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1106.651194] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b609a7-6249-d3c8-3cee-7ee8244cefa9" [ 1106.651194] env[62000]: _type = "HttpNfcLease" [ 1106.651194] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1106.651194] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1106.651194] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b609a7-6249-d3c8-3cee-7ee8244cefa9" [ 1106.651194] env[62000]: _type = "HttpNfcLease" [ 1106.651194] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1106.653584] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50bebaba-fcc8-4d76-87c7-a149639a55f9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.670614] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1106.670614] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating HTTP connection to write to file with size = 31668224 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1106.748531] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: a432ca35-942d-434a-9cc9-e6e0302a44fd] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1106.755868] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-2e5c8831-cc96-4816-9754-2705679c2724 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.834138] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.834490] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.834688] env[62000]: DEBUG nova.objects.instance [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'pci_requests' on Instance uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.838521] env[62000]: DEBUG nova.compute.manager [req-7a2801d9-98c0-4598-8c53-a4397a8e0067 req-2ed7766b-7f90-4957-8441-fe0cf9b59426 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Received event network-vif-deleted-8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1106.839086] env[62000]: INFO nova.compute.manager [req-7a2801d9-98c0-4598-8c53-a4397a8e0067 req-2ed7766b-7f90-4957-8441-fe0cf9b59426 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Neutron deleted interface 8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad; detaching it from the instance and deleting it from the info cache [ 1106.839086] env[62000]: DEBUG nova.network.neutron [req-7a2801d9-98c0-4598-8c53-a4397a8e0067 req-2ed7766b-7f90-4957-8441-fe0cf9b59426 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.252640] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 3780b22d-c360-4433-9f6b-9d5d1f14b525] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.317724] env[62000]: DEBUG nova.network.neutron [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.327432] env[62000]: DEBUG nova.network.neutron [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.338098] env[62000]: DEBUG nova.objects.instance [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'numa_topology' on Instance uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1107.342228] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d39cf4b0-52a6-48ff-a052-f49b3a3b352c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.354840] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bea2c4-11f2-4e47-aadd-0ebd2c9dcf33 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.386818] env[62000]: DEBUG nova.compute.manager [req-7a2801d9-98c0-4598-8c53-a4397a8e0067 req-2ed7766b-7f90-4957-8441-fe0cf9b59426 service nova] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Detach interface failed, port_id=8ab50541-4f1e-47ed-bfd6-540b8bd4c3ad, reason: Instance e0c78a4a-fe40-46cd-b0f3-dac971866cba could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1107.756077] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 4e3f804c-f4a2-44ab-bc84-3c97d5803fab] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1107.819834] env[62000]: INFO nova.compute.manager [-] [instance: e0c78a4a-fe40-46cd-b0f3-dac971866cba] Took 1.30 seconds to deallocate network for instance. [ 1107.829921] env[62000]: INFO nova.compute.manager [-] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Took 1.40 seconds to deallocate network for instance. [ 1107.841698] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1107.841907] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1107.842362] env[62000]: INFO nova.compute.claims [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1107.845325] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9be18e4b-e99b-45e3-b80a-09c8cd1c83eb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.852862] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1107.853082] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1107.853342] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-87c780a4-0dc3-4146-b816-271d9100c7d0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.021053] env[62000]: DEBUG oslo_vmware.rw_handles [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/524040ed-b1e5-7a24-8fa8-301a1998aa3f/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1108.021371] env[62000]: INFO nova.virt.vmwareapi.images [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Downloaded image file data 38a40578-3968-46a5-b445-a5b08f645b60 [ 1108.022323] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103e71b7-e199-423c-bc84-6f3e1167bb69 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.038179] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-851837ab-0478-4626-a20b-f6a8e8b2b9ae {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.068072] env[62000]: INFO nova.virt.vmwareapi.images [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] The imported VM was unregistered [ 1108.069828] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1108.070211] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Creating directory with path [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1108.070530] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-83eeb815-c80a-428c-8c16-bed4c50dbe3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.083904] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Created directory with path [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1108.084125] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77/OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77.vmdk to [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk. {{(pid=62000) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1108.084471] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-eadeb4a4-9575-4426-aae9-8989b18a7fc5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.090864] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1108.090864] env[62000]: value = "task-882963" [ 1108.090864] env[62000]: _type = "Task" [ 1108.090864] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.099224] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.259612] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: e9f2cd53-2c73-4c1f-858a-98a291da5547] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.327713] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.336994] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.600534] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.763222] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 8a11689f-fc00-43f8-9215-8d81daa84400] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1108.876853] env[62000]: DEBUG nova.compute.manager [req-f20f2233-80d6-4588-837e-d59dd50d6022 req-7ca02c5d-4bf2-4bcf-babd-84d1e92bcb05 service nova] [instance: c7d808d7-fa85-484c-822b-25f3cda5090b] Received event network-vif-deleted-cdfe8b13-6abd-43a6-a40e-aa5644dfe6d5 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1108.941151] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aa2876e-959a-4b26-baf9-0f4e7bcfc24b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.950575] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65fb2d4b-cf80-44d4-a932-d8b836dcb6b3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.988538] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247dae1a-88b5-4253-8eef-d0a40ce71e9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.996347] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfefc8a-b4de-423a-be2e-f5d3343bde52 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.010250] env[62000]: DEBUG nova.compute.provider_tree [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1109.101153] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.267143] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 0a3be3e8-b079-4006-8a46-9b9dd02baa5b] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1109.532140] env[62000]: ERROR nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [req-032054ad-015d-4598-b0da-2b91825c205c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-032054ad-015d-4598-b0da-2b91825c205c"}]} [ 1109.549077] env[62000]: DEBUG nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1109.564496] env[62000]: DEBUG nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1109.564757] env[62000]: DEBUG nova.compute.provider_tree [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1109.577088] env[62000]: DEBUG nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1109.597571] env[62000]: DEBUG nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1109.605445] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task} progress is 69%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.673696] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3938d0cb-d8a6-44a8-a930-fcfe0a27f597 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.681275] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc66fcf-33c4-4cad-ab58-11f874cd2992 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.713864] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87893c25-f570-40cb-b6ab-68d24557d27a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.722551] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8d104de-e7f0-47c8-8b6a-8913908b479c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.737070] env[62000]: DEBUG nova.compute.provider_tree [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1109.770829] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 48d29490-c174-4702-a449-3bf10bc9207c] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.102818] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task} progress is 91%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.267634] env[62000]: DEBUG nova.scheduler.client.report [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 146 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1110.267982] env[62000]: DEBUG nova.compute.provider_tree [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 146 to 147 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1110.268098] env[62000]: DEBUG nova.compute.provider_tree [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1110.275029] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 2f35dab3-6c32-4a35-91ca-1d6859e27f2c] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.603613] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882963, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.190585} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.603901] env[62000]: INFO nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77/OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77.vmdk to [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk. [ 1110.604113] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Cleaning up location [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1110.604288] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_26bdaa81-a2c7-4abf-a0a7-13f737fefd77 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1110.604541] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ea5deb0c-bc5e-4b1c-9b7a-280c19245142 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.610421] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1110.610421] env[62000]: value = "task-882964" [ 1110.610421] env[62000]: _type = "Task" [ 1110.610421] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.617652] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882964, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.774059] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.939s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.775866] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.448s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.776135] env[62000]: DEBUG nova.objects.instance [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lazy-loading 'resources' on Instance uuid e0c78a4a-fe40-46cd-b0f3-dac971866cba {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1110.777331] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 66edb906-f936-4f57-833d-224f36af109e] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1110.805455] env[62000]: INFO nova.network.neutron [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating port 69c33d12-7f23-4311-89f4-5b66d2fc837e with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1111.120120] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882964, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.036803} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.120373] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1111.120574] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Releasing lock "[datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1111.120828] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk to [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1111.121088] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6ec10b25-2ebf-419b-af2c-9632864cba16 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.127253] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1111.127253] env[62000]: value = "task-882965" [ 1111.127253] env[62000]: _type = "Task" [ 1111.127253] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.134420] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.283458] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: e2e3270d-086b-4441-a3d6-49b05a60b51f] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.366451] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3235ce18-2249-4b29-a49f-84d6b221e755 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.376444] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28d925e6-4b2f-45d7-9f5c-d07ed857e339 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.410169] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88d59c7-ff65-4ea4-9a0b-e1fc27ab683b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.418589] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fa77b5b-304d-413c-bd11-b457897f7ca7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.433279] env[62000]: DEBUG nova.compute.provider_tree [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1111.639138] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.786930] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 85f5f90a-9591-4393-9bef-2ad7e6a1a82f] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1111.963649] env[62000]: DEBUG nova.scheduler.client.report [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 147 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1111.963931] env[62000]: DEBUG nova.compute.provider_tree [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 147 to 148 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1111.964138] env[62000]: DEBUG nova.compute.provider_tree [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1112.138666] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.186307] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.186547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.235039] env[62000]: DEBUG nova.compute.manager [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1112.235299] env[62000]: DEBUG oslo_concurrency.lockutils [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1112.235510] env[62000]: DEBUG oslo_concurrency.lockutils [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.235683] env[62000]: DEBUG oslo_concurrency.lockutils [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.235861] env[62000]: DEBUG nova.compute.manager [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] No waiting events found dispatching network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1112.236234] env[62000]: WARNING nova.compute.manager [req-75de7400-1be5-4b0f-8297-7dd9517b2233 req-a6df2190-ad2a-448d-8c96-a14fe1eed98d service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received unexpected event network-vif-plugged-69c33d12-7f23-4311-89f4-5b66d2fc837e for instance with vm_state shelved_offloaded and task_state spawning. [ 1112.290449] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 96107e36-bf3f-4ef0-8d8b-5c9601f4f514] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1112.326034] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1112.326167] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1112.326328] env[62000]: DEBUG nova.network.neutron [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1112.468822] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.693s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1112.471168] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.134s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1112.471430] env[62000]: DEBUG nova.objects.instance [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lazy-loading 'resources' on Instance uuid c7d808d7-fa85-484c-822b-25f3cda5090b {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1112.492793] env[62000]: INFO nova.scheduler.client.report [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted allocations for instance e0c78a4a-fe40-46cd-b0f3-dac971866cba [ 1112.639367] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task} progress is 69%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.690418] env[62000]: DEBUG nova.compute.utils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1112.794115] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 153b86d9-dc2e-463d-87fd-155ec23e2abb] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.001749] env[62000]: DEBUG oslo_concurrency.lockutils [None req-a549f2c2-7047-486f-80fd-f54faa41d98e tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "e0c78a4a-fe40-46cd-b0f3-dac971866cba" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.600s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.041448] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a6c6222-41b5-40e5-b560-7392410a6a97 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.052205] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7b264b2-37af-4a3f-aa23-84b85a4e85c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.085926] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b08ecefd-1f86-44c1-8063-0f7217fc5d79 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.094276] env[62000]: DEBUG nova.network.neutron [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1113.096568] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7cd7b1d-b8b0-4846-825b-d99082944cc7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.111207] env[62000]: DEBUG nova.compute.provider_tree [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1113.139452] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task} progress is 91%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.193547] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1113.297201] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: ae1e8ebc-ddb8-4e95-847e-b9684e9161d4] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1113.602450] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.614107] env[62000]: DEBUG nova.scheduler.client.report [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1113.628380] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='75ffa1b81a0b699c6d6ea6b41d082ecc',container_format='bare',created_at=2024-09-16T07:02:23Z,direct_url=,disk_format='vmdk',id=74362b29-2b5c-4427-a86f-be14376f4ef3,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-1143947709-shelved',owner='4d131f95ad3949d89cd6f36f6648d3f1',properties=ImageMetaProps,protected=,size=31669760,status='active',tags=,updated_at=2024-09-16T07:02:37Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1113.628632] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1113.628793] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1113.628980] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1113.629147] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1113.629303] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1113.629513] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1113.629676] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1113.629847] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1113.630023] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1113.630210] env[62000]: DEBUG nova.virt.hardware [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1113.631047] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-306d3fe8-59ba-4185-98cf-e214e501f238 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.644946] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74d16f1-7748-4028-a8c2-ee51fb742256 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.648617] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882965, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.166276} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.649084] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/38a40578-3968-46a5-b445-a5b08f645b60/38a40578-3968-46a5-b445-a5b08f645b60.vmdk to [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1113.650105] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6699690c-b103-477e-99b0-911dc15eed69 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.659762] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ef:cc:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6a2e2e51-010f-4535-ba88-433663275996', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '69c33d12-7f23-4311-89f4-5b66d2fc837e', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1113.667057] env[62000]: DEBUG oslo.service.loopingcall [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1113.667558] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1113.667752] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d31b2afb-8bc1-4fe7-99c0-d38082d55a83 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.698298] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1113.698959] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d6ec38b-8909-4f50-ad5c-d4322a80587d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.714900] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1113.714900] env[62000]: value = "task-882966" [ 1113.714900] env[62000]: _type = "Task" [ 1113.714900] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.719748] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1113.719748] env[62000]: value = "task-882967" [ 1113.719748] env[62000]: _type = "Task" [ 1113.719748] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.722933] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882966, 'name': CreateVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.730903] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882967, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.800588] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: ae343199-1006-473a-a47f-7983835bd60a] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.120138] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.144174] env[62000]: INFO nova.scheduler.client.report [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted allocations for instance c7d808d7-fa85-484c-822b-25f3cda5090b [ 1114.224853] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882966, 'name': CreateVM_Task, 'duration_secs': 0.341116} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.227830] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1114.228528] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.228706] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.229086] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1114.229834] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-628a856e-9275-4fbd-9cfd-86707d53b0f4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.234686] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882967, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.237648] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1114.237648] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520256fa-68ab-db2d-d5f1-55c8245d212a" [ 1114.237648] env[62000]: _type = "Task" [ 1114.237648] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.245695] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]520256fa-68ab-db2d-d5f1-55c8245d212a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.255738] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.256013] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.256194] env[62000]: INFO nova.compute.manager [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Attaching volume b5778927-073f-41a2-b662-428d1e3d5fb7 to /dev/sdb [ 1114.259977] env[62000]: DEBUG nova.compute.manager [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1114.260191] env[62000]: DEBUG nova.compute.manager [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing instance network info cache due to event network-changed-69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1114.260405] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] Acquiring lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.260551] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] Acquired lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.260811] env[62000]: DEBUG nova.network.neutron [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Refreshing network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1114.291717] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1b12252-8036-4402-adab-abd43376268a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.298722] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5eee1356-f0e0-49cf-b336-b1c0c7d89f4a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.303776] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 296d5fea-e9d2-49e3-bc29-8e4bf3782535] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.311165] env[62000]: DEBUG nova.virt.block_device [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating existing volume attachment record: c2cfadc3-10ba-44d5-ad7c-069fa785cb79 {{(pid=62000) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1114.651374] env[62000]: DEBUG oslo_concurrency.lockutils [None req-af3bc6e9-aeed-4e8e-922e-ed25ac0ee2b0 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "c7d808d7-fa85-484c-822b-25f3cda5090b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.331s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.734301] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882967, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.747517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1114.747517] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Processing image 74362b29-2b5c-4427-a86f-be14376f4ef3 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1114.747517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1114.747517] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquired lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1114.747517] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1114.748133] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-23f5874e-4905-4821-bc1d-7df8ffbbebc2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.766434] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1114.766629] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1114.767386] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a0dc74b1-ac40-4cbf-a302-415b32433e5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.773155] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1114.773155] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c56294-bd65-ac40-2eb9-8f6c81baf774" [ 1114.773155] env[62000]: _type = "Task" [ 1114.773155] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1114.780714] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52c56294-bd65-ac40-2eb9-8f6c81baf774, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.806775] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 79a0d6f0-7dca-4c5d-8656-91bf4d957e78] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1114.983767] env[62000]: DEBUG nova.network.neutron [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updated VIF entry in instance network info cache for port 69c33d12-7f23-4311-89f4-5b66d2fc837e. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1114.984234] env[62000]: DEBUG nova.network.neutron [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [{"id": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "address": "fa:16:3e:ef:cc:85", "network": {"id": "414f23a8-825f-47fd-82f7-68df76378748", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1251917451-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d131f95ad3949d89cd6f36f6648d3f1", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6a2e2e51-010f-4535-ba88-433663275996", "external-id": "nsx-vlan-transportzone-915", "segmentation_id": 915, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap69c33d12-7f", "ovs_interfaceid": "69c33d12-7f23-4311-89f4-5b66d2fc837e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1115.234842] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882967, 'name': ReconfigVM_Task, 'duration_secs': 1.274376} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.235148] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6/15fde663-fac4-4198-962f-8f814f1317f6.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1115.235767] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7fbef53-e7e1-4214-86f3-1b787ef0d073 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.242372] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1115.242372] env[62000]: value = "task-882969" [ 1115.242372] env[62000]: _type = "Task" [ 1115.242372] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.250079] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882969, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.282941] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Preparing fetch location {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1115.283243] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Fetch image to [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef/OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef.vmdk {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1115.283449] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Downloading stream optimized image 74362b29-2b5c-4427-a86f-be14376f4ef3 to [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef/OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef.vmdk on the data store datastore1 as vApp {{(pid=62000) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1115.283666] env[62000]: DEBUG nova.virt.vmwareapi.images [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Downloading image file data 74362b29-2b5c-4427-a86f-be14376f4ef3 to the ESX as VM named 'OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef' {{(pid=62000) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1115.323395] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 2b0c7354-1622-4318-ba09-6e3214eed4db] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.360144] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1115.360144] env[62000]: value = "resgroup-9" [ 1115.360144] env[62000]: _type = "ResourcePool" [ 1115.360144] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1115.360469] env[62000]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-8ba8c3ff-945a-43ac-b994-22e22b4d1390 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.382056] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease: (returnval){ [ 1115.382056] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994640-6c96-ed88-e4b6-ac73be54a751" [ 1115.382056] env[62000]: _type = "HttpNfcLease" [ 1115.382056] env[62000]: } obtained for vApp import into resource pool (val){ [ 1115.382056] env[62000]: value = "resgroup-9" [ 1115.382056] env[62000]: _type = "ResourcePool" [ 1115.382056] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1115.382450] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the lease: (returnval){ [ 1115.382450] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994640-6c96-ed88-e4b6-ac73be54a751" [ 1115.382450] env[62000]: _type = "HttpNfcLease" [ 1115.382450] env[62000]: } to be ready. {{(pid=62000) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1115.388514] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.388514] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994640-6c96-ed88-e4b6-ac73be54a751" [ 1115.388514] env[62000]: _type = "HttpNfcLease" [ 1115.388514] env[62000]: } is initializing. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1115.487449] env[62000]: DEBUG oslo_concurrency.lockutils [req-e0b726ca-2991-467a-a217-c820faa9188e req-d580e002-75fa-44fa-98ad-226c0ebfd19b service nova] Releasing lock "refresh_cache-19410eb7-0fec-4270-89da-04a2975fc050" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1115.661740] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "fc04e119-c483-423c-8d94-d463eef6146e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.661953] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.689348] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1115.689580] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1115.752132] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882969, 'name': Rename_Task, 'duration_secs': 0.139128} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.752415] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1115.752666] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-57566b99-097f-4410-9ccb-ba55c6d18b22 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.758452] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1115.758452] env[62000]: value = "task-882971" [ 1115.758452] env[62000]: _type = "Task" [ 1115.758452] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.765860] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882971, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.826641] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 4c49fdd0-2485-4791-9349-a79a8663bbc0] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1115.890579] env[62000]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1115.890579] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994640-6c96-ed88-e4b6-ac73be54a751" [ 1115.890579] env[62000]: _type = "HttpNfcLease" [ 1115.890579] env[62000]: } is ready. {{(pid=62000) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1115.890895] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1115.890895] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52994640-6c96-ed88-e4b6-ac73be54a751" [ 1115.890895] env[62000]: _type = "HttpNfcLease" [ 1115.890895] env[62000]: }. {{(pid=62000) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1115.891642] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce01811c-1b46-43b9-a73a-46b97dd75270 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.900367] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk from lease info. {{(pid=62000) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1115.900572] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating HTTP connection to write to file with size = 31669760 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk. {{(pid=62000) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1115.966031] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ec9444f7-7cc9-43f9-ac04-f24d85297543 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.164228] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1116.191874] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1116.269547] env[62000]: DEBUG oslo_vmware.api [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-882971, 'name': PowerOnVM_Task, 'duration_secs': 0.465555} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.271488] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1116.330499] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: df525129-0ccb-4863-8a22-dd3e5a1aa2b5] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.384599] env[62000]: DEBUG nova.compute.manager [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1116.385769] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfcc056-dd45-4d18-aa0c-88b01b64e57d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1116.689135] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.689534] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.691296] env[62000]: INFO nova.compute.claims [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.712991] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.835605] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: ea8ae421-1d7f-4814-bc0f-90a3316ad028] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1116.910949] env[62000]: DEBUG oslo_concurrency.lockutils [None req-d69c663b-1c04-4e66-984a-b1c5c6571dde tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.364s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1117.125528] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Completed reading data from the image iterator. {{(pid=62000) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1117.125807] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1117.126930] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fa17a7a-af6a-4b63-ae6d-0b58004fb9bc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.135077] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk is in state: ready. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1117.135250] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk. {{(pid=62000) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1117.135518] env[62000]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-09d31d47-0466-41dd-b114-358f9450117a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.322373] env[62000]: DEBUG oslo_vmware.rw_handles [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/5234c70f-7652-e462-e9fc-ee10c25b1032/disk-0.vmdk. {{(pid=62000) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1117.322623] env[62000]: INFO nova.virt.vmwareapi.images [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Downloaded image file data 74362b29-2b5c-4427-a86f-be14376f4ef3 [ 1117.323516] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f917c1cd-ad87-4ac5-b2a8-edf3b6780386 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.338705] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 30d82c70-1401-4a1a-a88b-f798f8fbf96a] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.340337] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05e01344-9350-4113-97f9-49176eee0c85 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.373191] env[62000]: INFO nova.virt.vmwareapi.images [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] The imported VM was unregistered [ 1117.375584] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Caching image {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1117.375826] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Creating directory with path [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1117.376100] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1a1d7baf-8b66-4654-9d12-1b7077f35243 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.398637] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Created directory with path [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3 {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1117.398840] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef/OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef.vmdk to [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk. {{(pid=62000) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1117.399097] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-c9ed915f-88f9-4746-90e1-7465c746a722 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.405566] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1117.405566] env[62000]: value = "task-882974" [ 1117.405566] env[62000]: _type = "Task" [ 1117.405566] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1117.414289] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1117.768017] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67caf10-16e9-4337-838b-82676f239ab4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.776478] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e038d7-97e6-45b3-bd39-28f8e91377d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.805510] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd47d816-444f-4a8f-a298-e9b7517710a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.812131] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-347cc66e-461f-44fe-9daf-64d139d84726 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.824772] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1117.842767] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f3a6527c-777e-4a07-9482-598de15d4eb3] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1117.914944] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.345356] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 71192360-6ee6-4876-bf37-da987a09cbb2] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.348510] env[62000]: ERROR nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [req-ce925ef7-a250-4a9f-b07f-ad5c03b7a8da] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID bcfccf20-49dd-4b91-819e-4373e67bf5ec. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ce925ef7-a250-4a9f-b07f-ad5c03b7a8da"}]} [ 1118.367040] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Refreshing inventories for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1118.382529] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating ProviderTree inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1118.382791] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 52, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1118.396926] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Refreshing aggregate associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, aggregates: None {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1118.421414] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.422924] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Refreshing trait associations for resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec, traits: COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK {{(pid=62000) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1118.503613] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f968399-ed9e-4ff7-9ce0-19490879ccd7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.513294] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfc76fe-891c-424c-af6e-87f371e93add {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.545047] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d3e76b5-c39e-460b-bfa1-f0a8899157d4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.553682] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59384664-9047-4c98-8875-c4fda28e13a4 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.568562] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1118.852863] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: f10356f8-5eb4-4f6a-bd5b-7fcfdce03e5c] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1118.857148] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Volume attach. Driver type: vmdk {{(pid=62000) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1118.857678] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201663', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'name': 'volume-b5778927-073f-41a2-b662-428d1e3d5fb7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58b4c1b2-11b2-4d52-91c6-6df943b500f3', 'attached_at': '', 'detached_at': '', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'serial': 'b5778927-073f-41a2-b662-428d1e3d5fb7'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1118.859062] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-546d9bdb-6351-4e93-9aa6-588829be0000 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.882733] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86b7900e-5f24-4ab7-8fb1-70a2902b1da6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.915223] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfiguring VM instance instance-0000006a to attach disk [datastore1] volume-b5778927-073f-41a2-b662-428d1e3d5fb7/volume-b5778927-073f-41a2-b662-428d1e3d5fb7.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1118.916224] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-05509530-6067-4d19-a039-e8e07c2e9aef {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1118.945150] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1118.946801] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1118.946801] env[62000]: value = "task-882975" [ 1118.946801] env[62000]: _type = "Task" [ 1118.946801] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1118.955644] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882975, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.104136] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updated inventory for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with generation 150 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 1119.104568] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating resource provider bcfccf20-49dd-4b91-819e-4373e67bf5ec generation from 150 to 151 during operation: update_inventory {{(pid=62000) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 1119.104773] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Updating inventory in ProviderTree for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1119.365452] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 210af329-4cdb-4c3f-9e82-e72a2ea79421] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.426038] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.458480] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882975, 'name': ReconfigVM_Task, 'duration_secs': 0.3833} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.458852] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfigured VM instance instance-0000006a to attach disk [datastore1] volume-b5778927-073f-41a2-b662-428d1e3d5fb7/volume-b5778927-073f-41a2-b662-428d1e3d5fb7.vmdk or device None with type thin {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1119.464687] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ec3c43df-6494-4238-a164-79faa06dd2b1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.481703] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1119.481703] env[62000]: value = "task-882976" [ 1119.481703] env[62000]: _type = "Task" [ 1119.481703] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.491638] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882976, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.610689] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.921s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.611358] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1119.614954] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.902s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.616672] env[62000]: INFO nova.compute.claims [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1119.869362] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 334fd514-c7b5-402f-935e-1d95f9b5dbc8] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1119.925877] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 74%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.992909] env[62000]: DEBUG oslo_vmware.api [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882976, 'name': ReconfigVM_Task, 'duration_secs': 0.144002} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.993354] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201663', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'name': 'volume-b5778927-073f-41a2-b662-428d1e3d5fb7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58b4c1b2-11b2-4d52-91c6-6df943b500f3', 'attached_at': '', 'detached_at': '', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'serial': 'b5778927-073f-41a2-b662-428d1e3d5fb7'} {{(pid=62000) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1120.122339] env[62000]: DEBUG nova.compute.utils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1120.126851] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1120.126994] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1120.189016] env[62000]: DEBUG nova.policy [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b04f86982584d59b613b508f99d2c12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15b98199f74fdeb39abb58210f1c0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1120.375233] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 5a8cded8-bcfb-4488-a736-fb6b6aad5a94] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.427305] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1120.495760] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Successfully created port: 020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1120.628388] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1120.709981] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a59e310a-91b1-4da1-9ae0-e928c9c9135c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.718133] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9277b85e-d916-4b76-9b82-552e38e25fa0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.749626] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b65195-533a-44d2-81e6-cf07f3f1b93a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.757796] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bca223b-4742-4f8a-96ef-32d9ac6eb466 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.771718] env[62000]: DEBUG nova.compute.provider_tree [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.877802] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 49f6c198-13b7-4c07-81d8-c010e7b0598e] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1120.926171] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.037621] env[62000]: DEBUG nova.objects.instance [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 58b4c1b2-11b2-4d52-91c6-6df943b500f3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1121.274428] env[62000]: DEBUG nova.scheduler.client.report [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.381400] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: eb27703f-b657-423a-90a9-a7c024a2e473] Instance has had 0 of 5 cleanup attempts {{(pid=62000) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11239}} [ 1121.427909] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882974, 'name': MoveVirtualDisk_Task, 'duration_secs': 3.743425} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.428242] env[62000]: INFO nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef/OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef.vmdk to [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk. [ 1121.428511] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Cleaning up location [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1121.428748] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_a20dc143-5565-4453-b917-30bbe37bc9ef {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1121.429061] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbe68127-1b46-4078-bef8-1114de046e0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.435226] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1121.435226] env[62000]: value = "task-882977" [ 1121.435226] env[62000]: _type = "Task" [ 1121.435226] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.442672] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882977, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.542659] env[62000]: DEBUG oslo_concurrency.lockutils [None req-70eb444d-102f-4ec9-9b24-cd3b57716e86 tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.287s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.637527] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1121.661722] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1121.661983] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1121.662200] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1121.662409] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1121.662567] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1121.662720] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1121.662937] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1121.663122] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1121.663310] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1121.663480] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1121.663656] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1121.664503] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48677344-d4c5-4916-91b8-31551f1298ab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.672114] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-332e4004-29f3-4a60-925f-14ee1e3e9946 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.779098] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.164s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.779635] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1121.824964] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.825258] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.898983] env[62000]: DEBUG nova.compute.manager [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Received event network-vif-plugged-020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1121.899242] env[62000]: DEBUG oslo_concurrency.lockutils [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] Acquiring lock "fc04e119-c483-423c-8d94-d463eef6146e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1121.899465] env[62000]: DEBUG oslo_concurrency.lockutils [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] Lock "fc04e119-c483-423c-8d94-d463eef6146e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1121.899638] env[62000]: DEBUG oslo_concurrency.lockutils [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] Lock "fc04e119-c483-423c-8d94-d463eef6146e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.899813] env[62000]: DEBUG nova.compute.manager [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] No waiting events found dispatching network-vif-plugged-020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1121.899983] env[62000]: WARNING nova.compute.manager [req-3e339d45-1c4c-4ef8-afa5-9468693163c4 req-e33b941c-2d7d-44d3-a75b-29b7488a99dc service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Received unexpected event network-vif-plugged-020d8f4e-255d-41c2-ae36-a2367af448af for instance with vm_state building and task_state spawning. [ 1121.945241] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882977, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.249844} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1121.945498] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1121.945665] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Releasing lock "[datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1121.945909] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk to [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1121.946174] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cbed9b00-5fce-4955-9593-fdd2a3d04b24 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.952818] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1121.952818] env[62000]: value = "task-882978" [ 1121.952818] env[62000]: _type = "Task" [ 1121.952818] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.960203] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1121.991757] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Successfully updated port: 020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1122.284915] env[62000]: DEBUG nova.compute.utils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1122.286477] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1122.286709] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1122.325123] env[62000]: DEBUG nova.policy [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b04f86982584d59b613b508f99d2c12', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'aa15b98199f74fdeb39abb58210f1c0e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1122.327259] env[62000]: INFO nova.compute.manager [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Detaching volume b5778927-073f-41a2-b662-428d1e3d5fb7 [ 1122.360465] env[62000]: INFO nova.virt.block_device [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Attempting to driver detach volume b5778927-073f-41a2-b662-428d1e3d5fb7 from mountpoint /dev/sdb [ 1122.360734] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Volume detach. Driver type: vmdk {{(pid=62000) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1122.360954] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201663', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'name': 'volume-b5778927-073f-41a2-b662-428d1e3d5fb7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58b4c1b2-11b2-4d52-91c6-6df943b500f3', 'attached_at': '', 'detached_at': '', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'serial': 'b5778927-073f-41a2-b662-428d1e3d5fb7'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1122.362286] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f97da38f-130e-4d40-936d-d79740cdddb2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.385832] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ddde5d-b8d8-4298-a8e6-fd4fa611df2b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.393021] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50849288-450e-4be2-a499-33b51f185d66 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.412922] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdab8df7-4644-4dba-b2d1-4aa050cf7949 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.428790] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] The volume has not been displaced from its original location: [datastore1] volume-b5778927-073f-41a2-b662-428d1e3d5fb7/volume-b5778927-073f-41a2-b662-428d1e3d5fb7.vmdk. No consolidation needed. {{(pid=62000) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1122.434167] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfiguring VM instance instance-0000006a to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1122.435895] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0b80d8e3-3832-4866-aefb-35b0aff456c3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.453750] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1122.453750] env[62000]: value = "task-882979" [ 1122.453750] env[62000]: _type = "Task" [ 1122.453750] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1122.465451] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.468565] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882979, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.494577] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.494752] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.494887] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1122.652200] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Successfully created port: 1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1122.790185] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1122.966714] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 32%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.969797] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882979, 'name': ReconfigVM_Task, 'duration_secs': 0.262735} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.970108] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Reconfigured VM instance instance-0000006a to detach disk 2001 {{(pid=62000) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1122.974870] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f206182a-66ff-43a4-89fb-2cc03483e22b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.990745] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1122.990745] env[62000]: value = "task-882980" [ 1122.990745] env[62000]: _type = "Task" [ 1122.990745] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.000941] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.030802] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1123.173356] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Updating instance_info_cache with network_info: [{"id": "020d8f4e-255d-41c2-ae36-a2367af448af", "address": "fa:16:3e:12:af:7c", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap020d8f4e-25", "ovs_interfaceid": "020d8f4e-255d-41c2-ae36-a2367af448af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1123.465235] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 52%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.500665] env[62000]: DEBUG oslo_vmware.api [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882980, 'name': ReconfigVM_Task, 'duration_secs': 0.147367} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.500986] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-201663', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'name': 'volume-b5778927-073f-41a2-b662-428d1e3d5fb7', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '58b4c1b2-11b2-4d52-91c6-6df943b500f3', 'attached_at': '', 'detached_at': '', 'volume_id': 'b5778927-073f-41a2-b662-428d1e3d5fb7', 'serial': 'b5778927-073f-41a2-b662-428d1e3d5fb7'} {{(pid=62000) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1123.676726] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1123.677044] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Instance network_info: |[{"id": "020d8f4e-255d-41c2-ae36-a2367af448af", "address": "fa:16:3e:12:af:7c", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap020d8f4e-25", "ovs_interfaceid": "020d8f4e-255d-41c2-ae36-a2367af448af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1123.677523] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:af:7c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089a7624-43ba-4fce-bfc0-63e4bb7f9aeb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '020d8f4e-255d-41c2-ae36-a2367af448af', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1123.685622] env[62000]: DEBUG oslo.service.loopingcall [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1123.685884] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1123.686138] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f003ef10-ed16-477e-8111-4bf923437b22 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.707398] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1123.707398] env[62000]: value = "task-882981" [ 1123.707398] env[62000]: _type = "Task" [ 1123.707398] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.715726] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882981, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.800349] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1123.830639] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.830913] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.831098] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.831290] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.831441] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.831591] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.831806] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.831967] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.832222] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.832418] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.832598] env[62000]: DEBUG nova.virt.hardware [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.833617] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9029b9c-6e8a-4774-bb22-3934d34bdc21 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.842949] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cadb3e76-7ded-40f6-90e4-9542b3856e32 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.878626] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1123.925010] env[62000]: DEBUG nova.compute.manager [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Received event network-changed-020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1123.925358] env[62000]: DEBUG nova.compute.manager [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Refreshing instance network info cache due to event network-changed-020d8f4e-255d-41c2-ae36-a2367af448af. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1123.925703] env[62000]: DEBUG oslo_concurrency.lockutils [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] Acquiring lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1123.925901] env[62000]: DEBUG oslo_concurrency.lockutils [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] Acquired lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1123.926360] env[62000]: DEBUG nova.network.neutron [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Refreshing network info cache for port 020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1123.968547] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 74%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.057531] env[62000]: DEBUG nova.objects.instance [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'flavor' on Instance uuid 58b4c1b2-11b2-4d52-91c6-6df943b500f3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1124.197219] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Successfully updated port: 1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1124.218727] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882981, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.384340] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._sync_power_states {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1124.467177] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task} progress is 97%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.629153] env[62000]: DEBUG nova.network.neutron [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Updated VIF entry in instance network info cache for port 020d8f4e-255d-41c2-ae36-a2367af448af. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1124.629516] env[62000]: DEBUG nova.network.neutron [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Updating instance_info_cache with network_info: [{"id": "020d8f4e-255d-41c2-ae36-a2367af448af", "address": "fa:16:3e:12:af:7c", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap020d8f4e-25", "ovs_interfaceid": "020d8f4e-255d-41c2-ae36-a2367af448af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.699692] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.699834] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.699937] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1124.718498] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882981, 'name': CreateVM_Task} progress is 99%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.890235] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Getting list of instances from cluster (obj){ [ 1124.890235] env[62000]: value = "domain-c8" [ 1124.890235] env[62000]: _type = "ClusterComputeResource" [ 1124.890235] env[62000]: } {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 1124.890746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5330d47-8beb-41d2-a7f5-d51344e3f0d7 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.904500] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Got total of 3 instances {{(pid=62000) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 1124.904658] env[62000]: WARNING nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] While synchronizing instance power states, found 5 instances in the database and 3 instances on the hypervisor. [ 1124.904796] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Triggering sync for uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.904986] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Triggering sync for uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.905154] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Triggering sync for uuid 58b4c1b2-11b2-4d52-91c6-6df943b500f3 {{(pid=62000) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.905305] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Triggering sync for uuid fc04e119-c483-423c-8d94-d463eef6146e {{(pid=62000) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.905453] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Triggering sync for uuid 9c462295-7d2e-42db-a9b7-6b0a17b254eb {{(pid=62000) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10336}} [ 1124.905776] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.905993] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "15fde663-fac4-4198-962f-8f814f1317f6" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.906273] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.906493] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.906700] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "fc04e119-c483-423c-8d94-d463eef6146e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.906901] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.907681] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6e1e22-1340-4a20-adf7-c9bdc291b9ac {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.965250] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882978, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.561874} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.965510] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/74362b29-2b5c-4427-a86f-be14376f4ef3/74362b29-2b5c-4427-a86f-be14376f4ef3.vmdk to [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1124.966279] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffa00e39-6778-4e4a-8a07-dbbd911826d1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.987900] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.988157] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d92bef09-d3a9-4647-a322-4bb1d8d15dfb {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.008261] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1125.008261] env[62000]: value = "task-882982" [ 1125.008261] env[62000]: _type = "Task" [ 1125.008261] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.015571] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882982, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.067754] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be03c4a8-0656-49f4-b377-aee373daf30a tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.242s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.068988] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.162s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.069978] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2adf838a-407a-4120-8d22-b41bc27f7d9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.132258] env[62000]: DEBUG oslo_concurrency.lockutils [req-d22e310e-e9ab-4195-8270-1fbc9151cb1e req-b6106e66-43ea-47fc-82b2-9bd628d70cca service nova] Releasing lock "refresh_cache-fc04e119-c483-423c-8d94-d463eef6146e" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.219194] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882981, 'name': CreateVM_Task, 'duration_secs': 1.385565} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.219378] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1125.220031] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.220220] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.220554] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.220817] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ef78254-e797-4d4f-b04a-296f61662a26 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.225356] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1125.225356] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b371c2-7646-653c-18b1-7043ecb286ab" [ 1125.225356] env[62000]: _type = "Task" [ 1125.225356] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.232836] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b371c2-7646-653c-18b1-7043ecb286ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.238306] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1125.363843] env[62000]: DEBUG nova.network.neutron [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Updating instance_info_cache with network_info: [{"id": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "address": "fa:16:3e:6e:f9:1e", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa07005-b5", "ovs_interfaceid": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.416126] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "15fde663-fac4-4198-962f-8f814f1317f6" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.520370] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882982, 'name': ReconfigVM_Task, 'duration_secs': 0.286392} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.520661] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050/19410eb7-0fec-4270-89da-04a2975fc050.vmdk or device None with type streamOptimized {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.521281] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1841947f-3259-40f2-93d7-4c39f91c6e00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.528377] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1125.528377] env[62000]: value = "task-882983" [ 1125.528377] env[62000]: _type = "Task" [ 1125.528377] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.535939] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882983, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.579376] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.510s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.735413] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52b371c2-7646-653c-18b1-7043ecb286ab, 'name': SearchDatastore_Task, 'duration_secs': 0.053927} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.735686] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.735926] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1125.736187] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.736343] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.736526] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1125.736778] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-899807f0-4f0b-47fe-8760-e774a0fa1c41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.750277] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1125.750465] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1125.751131] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7392ec0-da49-4ea5-9c0d-5ff1da64ffe1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.755742] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1125.755742] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526c5228-bcef-2b0a-5cd4-86ff6b09523d" [ 1125.755742] env[62000]: _type = "Task" [ 1125.755742] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.763069] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526c5228-bcef-2b0a-5cd4-86ff6b09523d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.866707] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.866984] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Instance network_info: |[{"id": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "address": "fa:16:3e:6e:f9:1e", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa07005-b5", "ovs_interfaceid": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1125.867407] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:f9:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '089a7624-43ba-4fce-bfc0-63e4bb7f9aeb', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1aa07005-b575-4ff2-b93f-f1f340e032e0', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1125.875179] env[62000]: DEBUG oslo.service.loopingcall [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.875427] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1125.875651] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d620674c-90d0-4fcc-933f-8c089fd338f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.895509] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1125.895509] env[62000]: value = "task-882984" [ 1125.895509] env[62000]: _type = "Task" [ 1125.895509] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.902470] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882984, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.953232] env[62000]: DEBUG nova.compute.manager [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Received event network-vif-plugged-1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.953480] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Acquiring lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.953729] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.953892] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1125.954078] env[62000]: DEBUG nova.compute.manager [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] No waiting events found dispatching network-vif-plugged-1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1125.954273] env[62000]: WARNING nova.compute.manager [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Received unexpected event network-vif-plugged-1aa07005-b575-4ff2-b93f-f1f340e032e0 for instance with vm_state building and task_state spawning. [ 1125.954443] env[62000]: DEBUG nova.compute.manager [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Received event network-changed-1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1125.954616] env[62000]: DEBUG nova.compute.manager [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Refreshing instance network info cache due to event network-changed-1aa07005-b575-4ff2-b93f-f1f340e032e0. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1125.954809] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Acquiring lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.954948] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Acquired lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.955142] env[62000]: DEBUG nova.network.neutron [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Refreshing network info cache for port 1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1126.037965] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882983, 'name': Rename_Task, 'duration_secs': 0.133961} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.038361] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1126.038640] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fb1abc2a-1941-4e0e-94d9-4d44a947b669 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.044865] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1126.044865] env[62000]: value = "task-882985" [ 1126.044865] env[62000]: _type = "Task" [ 1126.044865] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.052383] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.083092] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.083520] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.083676] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1126.083879] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1126.084083] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.086405] env[62000]: INFO nova.compute.manager [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Terminating instance [ 1126.088260] env[62000]: DEBUG nova.compute.manager [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1126.088504] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1126.089409] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e41b3203-386d-4702-9c95-d180ee2c1a95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.096957] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1126.097225] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-182ac613-ebd5-4aca-8d3b-57b64a579c0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.103238] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1126.103238] env[62000]: value = "task-882986" [ 1126.103238] env[62000]: _type = "Task" [ 1126.103238] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.111730] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.268531] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]526c5228-bcef-2b0a-5cd4-86ff6b09523d, 'name': SearchDatastore_Task, 'duration_secs': 0.010126} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.269282] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5372a015-4b47-4cad-afc0-86c7c5ba0d33 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.275210] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1126.275210] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ef490a-a6c7-4fa0-3536-6bb778e609ee" [ 1126.275210] env[62000]: _type = "Task" [ 1126.275210] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.283250] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ef490a-a6c7-4fa0-3536-6bb778e609ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.405470] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-882984, 'name': CreateVM_Task, 'duration_secs': 0.379881} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.405635] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1126.406322] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.406498] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.406823] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1126.407099] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e354e835-f457-4d60-9a6f-7a53f4c2b020 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.411299] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1126.411299] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5261f59d-346b-48fb-306d-27c495c2a206" [ 1126.411299] env[62000]: _type = "Task" [ 1126.411299] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.419532] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5261f59d-346b-48fb-306d-27c495c2a206, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.553793] env[62000]: DEBUG oslo_vmware.api [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882985, 'name': PowerOnVM_Task, 'duration_secs': 0.495416} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.554156] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1126.612745] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882986, 'name': PowerOffVM_Task, 'duration_secs': 0.204196} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.613545] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1126.613545] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1126.613545] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-336c9c62-2672-4a49-b9ff-3732173d9c08 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.660704] env[62000]: DEBUG nova.compute.manager [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.661886] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b494abbc-1fc8-491f-a247-6ec0305955e3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.671414] env[62000]: DEBUG nova.network.neutron [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Updated VIF entry in instance network info cache for port 1aa07005-b575-4ff2-b93f-f1f340e032e0. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1126.671921] env[62000]: DEBUG nova.network.neutron [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Updating instance_info_cache with network_info: [{"id": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "address": "fa:16:3e:6e:f9:1e", "network": {"id": "8a88e9bb-51fb-482f-a3e6-3dbca1bfa808", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-12419835-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "aa15b98199f74fdeb39abb58210f1c0e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "089a7624-43ba-4fce-bfc0-63e4bb7f9aeb", "external-id": "nsx-vlan-transportzone-218", "segmentation_id": 218, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa07005-b5", "ovs_interfaceid": "1aa07005-b575-4ff2-b93f-f1f340e032e0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.786089] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52ef490a-a6c7-4fa0-3536-6bb778e609ee, 'name': SearchDatastore_Task, 'duration_secs': 0.00986} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.786089] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.786250] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] fc04e119-c483-423c-8d94-d463eef6146e/fc04e119-c483-423c-8d94-d463eef6146e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1126.786514] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-26489cde-953a-41c9-993d-63c08d9bbc6d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.793504] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1126.793504] env[62000]: value = "task-882988" [ 1126.793504] env[62000]: _type = "Task" [ 1126.793504] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.800942] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882988, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.921037] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5261f59d-346b-48fb-306d-27c495c2a206, 'name': SearchDatastore_Task, 'duration_secs': 0.009802} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.921352] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.921591] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1126.921827] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.921980] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.922248] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.922505] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-54713e5a-95d1-4c8b-9774-d0bf66b2b7cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.929335] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.929513] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1126.930176] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6e91d87-c4c7-475d-b8a3-6e11a87817ba {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.934975] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1126.934975] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f646ed-0276-f2eb-4595-83a219f6d545" [ 1126.934975] env[62000]: _type = "Task" [ 1126.934975] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.941873] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f646ed-0276-f2eb-4595-83a219f6d545, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.036733] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1127.036955] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1127.037273] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleting the datastore file [datastore1] 58b4c1b2-11b2-4d52-91c6-6df943b500f3 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1127.037586] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e3cc0272-21a2-4eeb-8e41-1b6ef429eae8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.044484] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for the task: (returnval){ [ 1127.044484] env[62000]: value = "task-882989" [ 1127.044484] env[62000]: _type = "Task" [ 1127.044484] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.052772] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882989, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.179237] env[62000]: DEBUG oslo_concurrency.lockutils [req-285fb028-cecc-4a12-8fd3-b097e3f1afb4 req-da296dfc-9f16-4732-9f7e-a4f8efb55727 service nova] Releasing lock "refresh_cache-9c462295-7d2e-42db-a9b7-6b0a17b254eb" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.182029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-52b34de7-5813-43e7-a556-e26bf7f1fae5 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 21.376s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.184029] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "19410eb7-0fec-4270-89da-04a2975fc050" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 2.277s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.184302] env[62000]: INFO nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] During sync_power_state the instance has a pending task (spawning). Skip. [ 1127.184572] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "19410eb7-0fec-4270-89da-04a2975fc050" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.303793] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882988, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.447209} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.304106] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] fc04e119-c483-423c-8d94-d463eef6146e/fc04e119-c483-423c-8d94-d463eef6146e.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1127.304328] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1127.304576] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2efac5d5-8bb0-4111-b9cb-4b00148315f2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.310465] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1127.310465] env[62000]: value = "task-882990" [ 1127.310465] env[62000]: _type = "Task" [ 1127.310465] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.318912] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882990, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.446164] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52f646ed-0276-f2eb-4595-83a219f6d545, 'name': SearchDatastore_Task, 'duration_secs': 0.00871} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.446932] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a7687a8-7399-4101-bfa4-e510400d9f9f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.453714] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1127.453714] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237142e-c9bb-e4af-0340-811db5726d54" [ 1127.453714] env[62000]: _type = "Task" [ 1127.453714] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.460847] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237142e-c9bb-e4af-0340-811db5726d54, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.554259] env[62000]: DEBUG oslo_vmware.api [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Task: {'id': task-882989, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.32049} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.554606] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1127.554693] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1127.554876] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1127.555075] env[62000]: INFO nova.compute.manager [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Took 1.47 seconds to destroy the instance on the hypervisor. [ 1127.555333] env[62000]: DEBUG oslo.service.loopingcall [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1127.555531] env[62000]: DEBUG nova.compute.manager [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1127.555644] env[62000]: DEBUG nova.network.neutron [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1127.820469] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882990, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069755} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.820767] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.821456] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1045589-71ed-4311-b3d3-d0ae69361e17 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.843222] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] fc04e119-c483-423c-8d94-d463eef6146e/fc04e119-c483-423c-8d94-d463eef6146e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.843495] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d3a756c3-aa40-4d0e-b04b-0ae27329e16a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.863207] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1127.863207] env[62000]: value = "task-882991" [ 1127.863207] env[62000]: _type = "Task" [ 1127.863207] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.871337] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882991, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.964169] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237142e-c9bb-e4af-0340-811db5726d54, 'name': SearchDatastore_Task, 'duration_secs': 0.009529} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.964468] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1127.964738] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 9c462295-7d2e-42db-a9b7-6b0a17b254eb/9c462295-7d2e-42db-a9b7-6b0a17b254eb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1127.964995] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-420c2be0-907a-4cb5-8c21-794a1276322c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.971455] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1127.971455] env[62000]: value = "task-882992" [ 1127.971455] env[62000]: _type = "Task" [ 1127.971455] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.979185] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882992, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.096821] env[62000]: DEBUG nova.compute.manager [req-da6bb532-e671-4478-8a7c-6e0e8d05351b req-19168586-248e-4764-95c8-e003e1ebde7a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Received event network-vif-deleted-679d672f-6163-425c-b4d9-c74d7d7b2a3f {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1128.097058] env[62000]: INFO nova.compute.manager [req-da6bb532-e671-4478-8a7c-6e0e8d05351b req-19168586-248e-4764-95c8-e003e1ebde7a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Neutron deleted interface 679d672f-6163-425c-b4d9-c74d7d7b2a3f; detaching it from the instance and deleting it from the info cache [ 1128.097247] env[62000]: DEBUG nova.network.neutron [req-da6bb532-e671-4478-8a7c-6e0e8d05351b req-19168586-248e-4764-95c8-e003e1ebde7a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.374626] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882991, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.482948] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882992, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.468274} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.483311] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 9c462295-7d2e-42db-a9b7-6b0a17b254eb/9c462295-7d2e-42db-a9b7-6b0a17b254eb.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1128.483536] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1128.483781] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d89fc4e1-73e1-4c27-8b73-6ac6eb9b6b6e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.489603] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1128.489603] env[62000]: value = "task-882993" [ 1128.489603] env[62000]: _type = "Task" [ 1128.489603] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.497390] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882993, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.578284] env[62000]: DEBUG nova.network.neutron [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1128.600184] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c49efcd-d687-40d6-b094-cd7a28bb9c3f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.609515] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f2a0e64-0729-4a5e-9007-bf51e8131b8c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.626786] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.627076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.627289] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "19410eb7-0fec-4270-89da-04a2975fc050-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.627474] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.627643] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.635610] env[62000]: DEBUG nova.compute.manager [req-da6bb532-e671-4478-8a7c-6e0e8d05351b req-19168586-248e-4764-95c8-e003e1ebde7a service nova] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Detach interface failed, port_id=679d672f-6163-425c-b4d9-c74d7d7b2a3f, reason: Instance 58b4c1b2-11b2-4d52-91c6-6df943b500f3 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1128.636229] env[62000]: INFO nova.compute.manager [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Terminating instance [ 1128.638264] env[62000]: DEBUG nova.compute.manager [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1128.638462] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1128.639236] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f195355-9db0-4fb2-aa4f-4d4cc6b88de3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.646422] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1128.646658] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-61bb8be9-4578-45dc-ae80-4fc9086cc20a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.652677] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1128.652677] env[62000]: value = "task-882994" [ 1128.652677] env[62000]: _type = "Task" [ 1128.652677] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.660724] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882994, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.874134] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882991, 'name': ReconfigVM_Task, 'duration_secs': 0.64888} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.874460] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Reconfigured VM instance instance-0000006d to attach disk [datastore1] fc04e119-c483-423c-8d94-d463eef6146e/fc04e119-c483-423c-8d94-d463eef6146e.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1128.875100] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-482e7f5e-cd2f-49fe-82f1-e03ba97f1992 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.881198] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1128.881198] env[62000]: value = "task-882995" [ 1128.881198] env[62000]: _type = "Task" [ 1128.881198] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.888916] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882995, 'name': Rename_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.000166] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882993, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063966} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.000465] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1129.001291] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55612d43-7951-41fc-a7a3-367214556b09 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.026674] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 9c462295-7d2e-42db-a9b7-6b0a17b254eb/9c462295-7d2e-42db-a9b7-6b0a17b254eb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1129.026944] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2599e879-9654-4e42-b900-7cf855e1fa36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.045879] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1129.045879] env[62000]: value = "task-882996" [ 1129.045879] env[62000]: _type = "Task" [ 1129.045879] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.053628] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882996, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.081371] env[62000]: INFO nova.compute.manager [-] [instance: 58b4c1b2-11b2-4d52-91c6-6df943b500f3] Took 1.53 seconds to deallocate network for instance. [ 1129.162610] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882994, 'name': PowerOffVM_Task, 'duration_secs': 0.192619} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.162833] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1129.163014] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1129.163310] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-117baa1f-96ee-485b-af04-13112bf424e2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.263132] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1129.263442] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1129.263637] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleting the datastore file [datastore1] 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1129.263951] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-24d3dbde-f8a3-4dab-a2d2-438ae582d4f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.269740] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for the task: (returnval){ [ 1129.269740] env[62000]: value = "task-882998" [ 1129.269740] env[62000]: _type = "Task" [ 1129.269740] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.279071] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882998, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.390804] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882995, 'name': Rename_Task, 'duration_secs': 0.143059} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.391187] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1129.391496] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-741dd1ef-f58b-4378-a36c-97012ed87dd5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.397176] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1129.397176] env[62000]: value = "task-882999" [ 1129.397176] env[62000]: _type = "Task" [ 1129.397176] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.404641] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882999, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.556905] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882996, 'name': ReconfigVM_Task, 'duration_secs': 0.290747} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.557239] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 9c462295-7d2e-42db-a9b7-6b0a17b254eb/9c462295-7d2e-42db-a9b7-6b0a17b254eb.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1129.557920] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a2b3dddd-0e5f-4dd4-aa89-4abcacf0436c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.564779] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1129.564779] env[62000]: value = "task-883000" [ 1129.564779] env[62000]: _type = "Task" [ 1129.564779] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1129.572614] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883000, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1129.587756] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.588079] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.588281] env[62000]: DEBUG nova.objects.instance [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lazy-loading 'resources' on Instance uuid 58b4c1b2-11b2-4d52-91c6-6df943b500f3 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1129.779265] env[62000]: DEBUG oslo_vmware.api [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Task: {'id': task-882998, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15067} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1129.779526] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1129.779716] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1129.779895] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1129.780102] env[62000]: INFO nova.compute.manager [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1129.780354] env[62000]: DEBUG oslo.service.loopingcall [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1129.780552] env[62000]: DEBUG nova.compute.manager [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1129.780648] env[62000]: DEBUG nova.network.neutron [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1129.907189] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882999, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.074550] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883000, 'name': Rename_Task, 'duration_secs': 0.145898} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.074931] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1130.075306] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-90485a4d-24ee-4cf6-a5b3-5afff4cc6962 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.082314] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1130.082314] env[62000]: value = "task-883001" [ 1130.082314] env[62000]: _type = "Task" [ 1130.082314] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.092608] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883001, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1130.174147] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28599c3d-4bff-461a-9222-110dbefa7ef2 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.180275] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03adad35-21bf-49eb-ad34-86292bb86a84 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.217810] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f774482-9d42-43da-877b-c6ffae8af60c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.227271] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ece6dce-8bf5-46ad-bf95-727898dab99f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.234599] env[62000]: DEBUG nova.compute.manager [req-1b15acce-558f-48cc-83ed-a6c5b4c39c1b req-450b1fed-e29c-4e7b-ad29-97ccfda98997 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Received event network-vif-deleted-69c33d12-7f23-4311-89f4-5b66d2fc837e {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1130.234840] env[62000]: INFO nova.compute.manager [req-1b15acce-558f-48cc-83ed-a6c5b4c39c1b req-450b1fed-e29c-4e7b-ad29-97ccfda98997 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Neutron deleted interface 69c33d12-7f23-4311-89f4-5b66d2fc837e; detaching it from the instance and deleting it from the info cache [ 1130.235138] env[62000]: DEBUG nova.network.neutron [req-1b15acce-558f-48cc-83ed-a6c5b4c39c1b req-450b1fed-e29c-4e7b-ad29-97ccfda98997 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.245150] env[62000]: DEBUG nova.compute.provider_tree [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1130.407784] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-882999, 'name': PowerOnVM_Task, 'duration_secs': 0.684806} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.408090] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1130.408358] env[62000]: INFO nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Took 8.77 seconds to spawn the instance on the hypervisor. [ 1130.408552] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1130.409405] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42ab5a66-03c6-4e44-8235-484b86a483a8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.592222] env[62000]: DEBUG oslo_vmware.api [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883001, 'name': PowerOnVM_Task, 'duration_secs': 0.468408} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1130.592533] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1130.592735] env[62000]: INFO nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Took 6.79 seconds to spawn the instance on the hypervisor. [ 1130.592917] env[62000]: DEBUG nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1130.593701] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-966bb9e8-a5c8-4745-9cf9-947e241382f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.721742] env[62000]: DEBUG nova.network.neutron [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.737831] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-48c7c9ac-509d-4e5f-ab31-0c85dc2486c9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.747143] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41219a5-c2fb-4cac-82f4-b2374c78e1cd {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.758073] env[62000]: DEBUG nova.scheduler.client.report [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.778456] env[62000]: DEBUG nova.compute.manager [req-1b15acce-558f-48cc-83ed-a6c5b4c39c1b req-450b1fed-e29c-4e7b-ad29-97ccfda98997 service nova] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Detach interface failed, port_id=69c33d12-7f23-4311-89f4-5b66d2fc837e, reason: Instance 19410eb7-0fec-4270-89da-04a2975fc050 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1130.927618] env[62000]: INFO nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Took 14.26 seconds to build instance. [ 1131.110016] env[62000]: INFO nova.compute.manager [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Took 14.41 seconds to build instance. [ 1131.224840] env[62000]: INFO nova.compute.manager [-] [instance: 19410eb7-0fec-4270-89da-04a2975fc050] Took 1.44 seconds to deallocate network for instance. [ 1131.263025] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.675s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.286341] env[62000]: INFO nova.scheduler.client.report [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Deleted allocations for instance 58b4c1b2-11b2-4d52-91c6-6df943b500f3 [ 1131.429677] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.768s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.429948] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "fc04e119-c483-423c-8d94-d463eef6146e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.523s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.430206] env[62000]: INFO nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: fc04e119-c483-423c-8d94-d463eef6146e] During sync_power_state the instance has a pending task (spawning). Skip. [ 1131.430407] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "fc04e119-c483-423c-8d94-d463eef6146e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.612109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-854840d0-f847-4bbb-8ddf-738f6a9b5868 tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.922s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.612109] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 6.705s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.612109] env[62000]: INFO nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] During sync_power_state the instance has a pending task (spawning). Skip. [ 1131.612629] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1131.730654] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.730938] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1131.731222] env[62000]: DEBUG nova.objects.instance [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lazy-loading 'resources' on Instance uuid 19410eb7-0fec-4270-89da-04a2975fc050 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1131.793190] env[62000]: DEBUG oslo_concurrency.lockutils [None req-540b5037-e3d6-4eb2-b263-70281acc113d tempest-AttachVolumeNegativeTest-1086000320 tempest-AttachVolumeNegativeTest-1086000320-project-member] Lock "58b4c1b2-11b2-4d52-91c6-6df943b500f3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.710s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.292814] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6150ce5-09cb-4f88-adb7-c735b2d3a12f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.301092] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2998d8-8278-4392-a590-32c86b857d6c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.331942] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6320e919-ca1e-410f-9c04-eb2338242fa5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.339791] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f67edf-22dc-41ed-89c1-d8812e4220fa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.353012] env[62000]: DEBUG nova.compute.provider_tree [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1132.427015] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "fc04e119-c483-423c-8d94-d463eef6146e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.427311] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.427541] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "fc04e119-c483-423c-8d94-d463eef6146e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.427729] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.427901] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.430100] env[62000]: INFO nova.compute.manager [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Terminating instance [ 1132.431903] env[62000]: DEBUG nova.compute.manager [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1132.432128] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1132.432950] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a6da294-d20f-4a96-b808-4aee2bcd58ce {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.440630] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1132.441035] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6787e506-145b-4535-94be-110f835a1704 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.447990] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1132.447990] env[62000]: value = "task-883002" [ 1132.447990] env[62000]: _type = "Task" [ 1132.447990] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.456008] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883002, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.488685] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.488980] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.489226] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.489789] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.489789] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.491861] env[62000]: INFO nova.compute.manager [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Terminating instance [ 1132.493780] env[62000]: DEBUG nova.compute.manager [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1132.493980] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1132.495069] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaba0cf-ab35-444e-8533-8d637aeb73e5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.503594] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1132.503865] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6ee7e52-6c07-46d8-8991-6b594706d23d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.510872] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1132.510872] env[62000]: value = "task-883003" [ 1132.510872] env[62000]: _type = "Task" [ 1132.510872] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.521066] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883003, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.856122] env[62000]: DEBUG nova.scheduler.client.report [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.958089] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883002, 'name': PowerOffVM_Task, 'duration_secs': 0.180355} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.958369] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1132.958546] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1132.958794] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6efa5821-3109-4ce6-80c1-eb91c1021f7e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.018835] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1133.018985] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1133.019193] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleting the datastore file [datastore1] fc04e119-c483-423c-8d94-d463eef6146e {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.019437] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f397c160-0289-46be-88ae-648a641f4c4e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.023961] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883003, 'name': PowerOffVM_Task, 'duration_secs': 0.19452} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.024530] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1133.024727] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1133.024961] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e9230858-c4b5-4834-a9dd-b0a336668df8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.028112] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1133.028112] env[62000]: value = "task-883006" [ 1133.028112] env[62000]: _type = "Task" [ 1133.028112] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.035132] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883006, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.085714] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1133.085949] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1133.086170] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleting the datastore file [datastore1] 9c462295-7d2e-42db-a9b7-6b0a17b254eb {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1133.086449] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff988904-fb63-451c-bd56-8513299d359d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.092396] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for the task: (returnval){ [ 1133.092396] env[62000]: value = "task-883008" [ 1133.092396] env[62000]: _type = "Task" [ 1133.092396] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.099941] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883008, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.361649] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.387061] env[62000]: INFO nova.scheduler.client.report [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Deleted allocations for instance 19410eb7-0fec-4270-89da-04a2975fc050 [ 1133.539215] env[62000]: DEBUG oslo_vmware.api [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883006, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130479} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.539421] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.539601] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1133.539775] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1133.539948] env[62000]: INFO nova.compute.manager [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1133.540240] env[62000]: DEBUG oslo.service.loopingcall [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.540447] env[62000]: DEBUG nova.compute.manager [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1133.540544] env[62000]: DEBUG nova.network.neutron [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1133.601714] env[62000]: DEBUG oslo_vmware.api [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Task: {'id': task-883008, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.124957} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.602018] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1133.602239] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1133.602657] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1133.602925] env[62000]: INFO nova.compute.manager [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1133.603154] env[62000]: DEBUG oslo.service.loopingcall [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1133.603593] env[62000]: DEBUG nova.compute.manager [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1133.603710] env[62000]: DEBUG nova.network.neutron [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1133.894289] env[62000]: DEBUG oslo_concurrency.lockutils [None req-be5567ed-5cfc-4be1-b5f7-3aef3ef519b2 tempest-ServerActionsTestOtherB-1573004507 tempest-ServerActionsTestOtherB-1573004507-project-member] Lock "19410eb7-0fec-4270-89da-04a2975fc050" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.267s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1133.907515] env[62000]: DEBUG nova.compute.manager [req-ee2b1001-0cf6-4aa8-a973-aa1ef73a54db req-c869eb44-b02d-49e1-ba91-6f98e3b1169c service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Received event network-vif-deleted-020d8f4e-255d-41c2-ae36-a2367af448af {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1133.907658] env[62000]: INFO nova.compute.manager [req-ee2b1001-0cf6-4aa8-a973-aa1ef73a54db req-c869eb44-b02d-49e1-ba91-6f98e3b1169c service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Neutron deleted interface 020d8f4e-255d-41c2-ae36-a2367af448af; detaching it from the instance and deleting it from the info cache [ 1133.907779] env[62000]: DEBUG nova.network.neutron [req-ee2b1001-0cf6-4aa8-a973-aa1ef73a54db req-c869eb44-b02d-49e1-ba91-6f98e3b1169c service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.381964] env[62000]: DEBUG nova.network.neutron [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.387781] env[62000]: DEBUG nova.network.neutron [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1134.410471] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c1c2598f-7602-40e0-bfe0-92524d23ab12 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.420422] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15c5c779-5d73-476e-9a12-8894d469fd89 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.444271] env[62000]: DEBUG nova.compute.manager [req-ee2b1001-0cf6-4aa8-a973-aa1ef73a54db req-c869eb44-b02d-49e1-ba91-6f98e3b1169c service nova] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Detach interface failed, port_id=020d8f4e-255d-41c2-ae36-a2367af448af, reason: Instance fc04e119-c483-423c-8d94-d463eef6146e could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1134.885100] env[62000]: INFO nova.compute.manager [-] [instance: fc04e119-c483-423c-8d94-d463eef6146e] Took 1.34 seconds to deallocate network for instance. [ 1134.890249] env[62000]: INFO nova.compute.manager [-] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Took 1.29 seconds to deallocate network for instance. [ 1135.392076] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.392427] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1135.392665] env[62000]: DEBUG nova.objects.instance [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lazy-loading 'resources' on Instance uuid fc04e119-c483-423c-8d94-d463eef6146e {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.397843] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.956400] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0e1ee11-8fcd-4a4b-ad96-079329834184 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.964077] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7528b6c4-71b4-4b27-ac5d-0031567ce42b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.969998] env[62000]: DEBUG nova.compute.manager [req-a1a0fdfc-0e3e-425e-90c4-481c93a4cb14 req-c0559188-9a6c-4fb2-917f-905ecc243aa5 service nova] [instance: 9c462295-7d2e-42db-a9b7-6b0a17b254eb] Received event network-vif-deleted-1aa07005-b575-4ff2-b93f-f1f340e032e0 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1135.998227] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829a83c2-be7c-4db0-a806-3708657aed00 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.006746] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be44b0ee-100b-4ade-bf81-b195f77dcc4b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.020206] env[62000]: DEBUG nova.compute.provider_tree [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.524221] env[62000]: DEBUG nova.scheduler.client.report [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1137.029596] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.637s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.031988] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.634s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1137.033384] env[62000]: DEBUG nova.objects.instance [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lazy-loading 'resources' on Instance uuid 9c462295-7d2e-42db-a9b7-6b0a17b254eb {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1137.053154] env[62000]: INFO nova.scheduler.client.report [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted allocations for instance fc04e119-c483-423c-8d94-d463eef6146e [ 1137.563477] env[62000]: DEBUG oslo_concurrency.lockutils [None req-53d70acd-bf95-4244-a77d-846a38ea4fdd tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "fc04e119-c483-423c-8d94-d463eef6146e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.136s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.585802] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9427695b-0076-4497-aa0b-ee2120e7ef53 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.593709] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51672ff2-bfe5-4ca0-b5b8-a9e9fae818c8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.624175] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4f913a-1e9b-46ee-baf2-f6610b78da36 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.631162] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29b1acd-6e68-4dcf-be1b-b622d0265f1a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.643744] env[62000]: DEBUG nova.compute.provider_tree [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1138.146626] env[62000]: DEBUG nova.scheduler.client.report [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.652324] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.620s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.675806] env[62000]: INFO nova.scheduler.client.report [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Deleted allocations for instance 9c462295-7d2e-42db-a9b7-6b0a17b254eb [ 1139.185300] env[62000]: DEBUG oslo_concurrency.lockutils [None req-25f42fb3-0241-4bd8-ac16-3487b28290fb tempest-MultipleCreateTestJSON-1748387172 tempest-MultipleCreateTestJSON-1748387172-project-member] Lock "9c462295-7d2e-42db-a9b7-6b0a17b254eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.696s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.638491] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "352b2627-6e8e-43c0-b724-43072e28fa5c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.638809] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.141548] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Starting instance... {{(pid=62000) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1141.554256] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1141.662882] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.663185] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.664962] env[62000]: INFO nova.compute.claims [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1142.714511] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab6e224-154c-4969-baab-be430c5ba977 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.722031] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4da64798-0928-4952-a675-1aba39b81bcc {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.751266] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9cdc996-ee0b-4216-a681-c47ce9421fbe {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.759074] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab3e492-1877-4550-bcc7-df026efc3e7c {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.771778] env[62000]: DEBUG nova.compute.provider_tree [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1143.025232] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.029863] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1143.030983] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Starting heal instance info cache {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 1143.274731] env[62000]: DEBUG nova.scheduler.client.report [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1143.779651] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.116s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.780216] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Start building networks asynchronously for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1144.035784] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Didn't find any instances for network info cache update. {{(pid=62000) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 1144.035985] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.036195] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager.update_available_resource {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.285539] env[62000]: DEBUG nova.compute.utils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Using /dev/sd instead of None {{(pid=62000) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1144.286991] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Allocating IP information in the background. {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1144.287214] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] allocate_for_instance() {{(pid=62000) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 1144.336392] env[62000]: DEBUG nova.policy [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fa6049f31eba42838660675b7a057c30', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e30718d1687a4d03a1b0ed498dae3e4f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62000) authorize /opt/stack/nova/nova/policy.py:201}} [ 1144.539036] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.539269] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.539439] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.539596] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62000) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1144.540543] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24f04ea-fa1e-460e-bd1c-b50ec3b112f0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.549764] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2972b1bb-c3a3-4c09-adee-e8abe1e7658b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.567187] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc37852e-481b-40c2-a9af-132174cfb973 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.573025] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e50c97b-ef25-4a5d-993a-6bf1d4e10dab {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.603117] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181166MB free_disk=53GB free_vcpus=48 pci_devices=None {{(pid=62000) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1144.603343] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1144.603446] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1144.612785] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Successfully created port: dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1144.792450] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Start building block device mappings for instance. {{(pid=62000) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1145.628360] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 15fde663-fac4-4198-962f-8f814f1317f6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.628525] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Instance 352b2627-6e8e-43c0-b724-43072e28fa5c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62000) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1145.628706] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1145.628848] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=62000) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1145.663912] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c46708b-b67f-4a81-b10c-d21880eff493 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.673051] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8706f715-ba4f-45c7-9ff6-ed55e6396162 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.702015] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23177814-2cb9-4091-ad16-b6fd41203942 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.709102] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8190170e-46c3-4de5-abd9-f89ee0b56b0d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.721599] env[62000]: DEBUG nova.compute.provider_tree [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.801984] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Start spawning the instance on the hypervisor. {{(pid=62000) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1145.828658] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-09-16T06:51:27Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-09-16T06:51:10Z,direct_url=,disk_format='vmdk',id=37b135ac-1e3f-46c1-bb0e-ce7f0b69db72,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='29e4579cf2604191945dca831f024a21',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-09-16T06:51:10Z,virtual_size=,visibility=), allow threads: False {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1145.828915] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Flavor limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1145.829091] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Image limits 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1145.829285] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Flavor pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1145.829461] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Image pref 0:0:0 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1145.829644] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62000) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1145.829859] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1145.830054] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1145.830249] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Got 1 possible topologies {{(pid=62000) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1145.830418] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1145.830764] env[62000]: DEBUG nova.virt.hardware [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62000) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1145.831561] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1021fe9a-b9ce-44cb-8f6e-e1d90a65a05f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.838874] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7225ec32-dc29-443d-9194-b3f47787403a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.101866] env[62000]: DEBUG nova.compute.manager [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Received event network-vif-plugged-dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1146.102160] env[62000]: DEBUG oslo_concurrency.lockutils [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] Acquiring lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1146.102399] env[62000]: DEBUG oslo_concurrency.lockutils [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1146.102576] env[62000]: DEBUG oslo_concurrency.lockutils [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.102749] env[62000]: DEBUG nova.compute.manager [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] No waiting events found dispatching network-vif-plugged-dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1146.103032] env[62000]: WARNING nova.compute.manager [req-46787338-c36e-4a0a-a4b7-3e61663a21ea req-00eaba5b-5e0f-4232-a111-3fa524a6de8d service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Received unexpected event network-vif-plugged-dc74c58b-7dae-4321-b90e-10ea9a636007 for instance with vm_state building and task_state spawning. [ 1146.182742] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Successfully updated port: dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1146.225191] env[62000]: DEBUG nova.scheduler.client.report [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.685902] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1146.685902] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquired lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1146.685902] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Building network info cache for instance {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 1146.729191] env[62000]: DEBUG nova.compute.resource_tracker [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62000) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1146.729467] env[62000]: DEBUG oslo_concurrency.lockutils [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.126s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.217524] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Instance cache missing network info. {{(pid=62000) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 1147.334306] env[62000]: DEBUG nova.network.neutron [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Updating instance_info_cache with network_info: [{"id": "dc74c58b-7dae-4321-b90e-10ea9a636007", "address": "fa:16:3e:df:20:4b", "network": {"id": "df8b90e9-b2cc-49c0-9a38-44e6f7ba1e28", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1460912285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e30718d1687a4d03a1b0ed498dae3e4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc74c58b-7d", "ovs_interfaceid": "dc74c58b-7dae-4321-b90e-10ea9a636007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1147.723092] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.723351] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1147.723502] env[62000]: DEBUG nova.compute.manager [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62000) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 1147.837528] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Releasing lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1147.837852] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Instance network_info: |[{"id": "dc74c58b-7dae-4321-b90e-10ea9a636007", "address": "fa:16:3e:df:20:4b", "network": {"id": "df8b90e9-b2cc-49c0-9a38-44e6f7ba1e28", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1460912285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e30718d1687a4d03a1b0ed498dae3e4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc74c58b-7d", "ovs_interfaceid": "dc74c58b-7dae-4321-b90e-10ea9a636007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62000) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1147.838305] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:df:20:4b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '489b2441-7132-4942-8b61-49cf0ad4400e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dc74c58b-7dae-4321-b90e-10ea9a636007', 'vif_model': 'vmxnet3'}] {{(pid=62000) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1147.845628] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Creating folder: Project (e30718d1687a4d03a1b0ed498dae3e4f). Parent ref: group-v201431. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1147.845891] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fb0de351-b4ce-4dd7-949f-de10888a188a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.857643] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Created folder: Project (e30718d1687a4d03a1b0ed498dae3e4f) in parent group-v201431. [ 1147.857825] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Creating folder: Instances. Parent ref: group-v201667. {{(pid=62000) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 1147.858061] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-43300542-9961-4b61-b796-67a86196ea78 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.866712] env[62000]: INFO nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Created folder: Instances in parent group-v201667. [ 1147.866936] env[62000]: DEBUG oslo.service.loopingcall [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1147.867133] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Creating VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 1147.867327] env[62000]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8a3c71e9-acc9-4c3d-900b-e652a088aaf3 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.885414] env[62000]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1147.885414] env[62000]: value = "task-883014" [ 1147.885414] env[62000]: _type = "Task" [ 1147.885414] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.892290] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-883014, 'name': CreateVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.030597] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1148.128256] env[62000]: DEBUG nova.compute.manager [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Received event network-changed-dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1148.128305] env[62000]: DEBUG nova.compute.manager [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Refreshing instance network info cache due to event network-changed-dc74c58b-7dae-4321-b90e-10ea9a636007. {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 1148.128539] env[62000]: DEBUG oslo_concurrency.lockutils [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] Acquiring lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.128772] env[62000]: DEBUG oslo_concurrency.lockutils [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] Acquired lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.128985] env[62000]: DEBUG nova.network.neutron [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Refreshing network info cache for port dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 1148.396028] env[62000]: DEBUG oslo_vmware.api [-] Task: {'id': task-883014, 'name': CreateVM_Task, 'duration_secs': 0.278444} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.396345] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Created VM on the ESX host {{(pid=62000) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 1148.396827] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.396999] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.397342] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1148.397589] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0580ef8-74aa-4d91-8de7-0336e47a446f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.401769] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1148.401769] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ea4d8-d7dc-5c03-2304-1f6429b8b1a8" [ 1148.401769] env[62000]: _type = "Task" [ 1148.401769] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.408810] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ea4d8-d7dc-5c03-2304-1f6429b8b1a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.808392] env[62000]: DEBUG nova.network.neutron [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Updated VIF entry in instance network info cache for port dc74c58b-7dae-4321-b90e-10ea9a636007. {{(pid=62000) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 1148.808798] env[62000]: DEBUG nova.network.neutron [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Updating instance_info_cache with network_info: [{"id": "dc74c58b-7dae-4321-b90e-10ea9a636007", "address": "fa:16:3e:df:20:4b", "network": {"id": "df8b90e9-b2cc-49c0-9a38-44e6f7ba1e28", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-1460912285-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e30718d1687a4d03a1b0ed498dae3e4f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "489b2441-7132-4942-8b61-49cf0ad4400e", "external-id": "nsx-vlan-transportzone-971", "segmentation_id": 971, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdc74c58b-7d", "ovs_interfaceid": "dc74c58b-7dae-4321-b90e-10ea9a636007", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.911290] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]521ea4d8-d7dc-5c03-2304-1f6429b8b1a8, 'name': SearchDatastore_Task, 'duration_secs': 0.010922} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.911543] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1148.911784] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Processing image 37b135ac-1e3f-46c1-bb0e-ce7f0b69db72 {{(pid=62000) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1148.912030] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.912217] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquired lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1148.912406] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1148.912665] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-43cad39a-4d0a-47a4-8a84-771780ebbb41 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.920258] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62000) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1148.920435] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62000) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 1148.921118] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f3df6e3e-79bd-44e9-bd1c-8c8cbccb2d95 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.926121] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1148.926121] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237c21d-1363-82d1-319e-379ff7a0c924" [ 1148.926121] env[62000]: _type = "Task" [ 1148.926121] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.932988] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237c21d-1363-82d1-319e-379ff7a0c924, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.030265] env[62000]: DEBUG oslo_service.periodic_task [None req-ad37f9b3-ad10-4c5b-8f54-879a97f9941d None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62000) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1149.311482] env[62000]: DEBUG oslo_concurrency.lockutils [req-e36c83cf-d6a5-4cde-823f-9d3edb49e1b2 req-50e69475-bf12-4808-8e0d-7a1c436a84be service nova] Releasing lock "refresh_cache-352b2627-6e8e-43c0-b724-43072e28fa5c" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.437140] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]5237c21d-1363-82d1-319e-379ff7a0c924, 'name': SearchDatastore_Task, 'duration_secs': 0.008085} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.437900] env[62000]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5d47654-0f48-44ff-82b3-74a5915c668f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.442646] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1149.442646] env[62000]: value = "session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dcb767-1a56-5e37-8509-82026c6f514b" [ 1149.442646] env[62000]: _type = "Task" [ 1149.442646] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.449837] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dcb767-1a56-5e37-8509-82026c6f514b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.954422] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': session[5269d046-f6ea-490a-8bbc-4a06cc908b5c]52dcb767-1a56-5e37-8509-82026c6f514b, 'name': SearchDatastore_Task, 'duration_secs': 0.008782} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.954677] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Releasing lock "[datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk" {{(pid=62000) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.954934] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 352b2627-6e8e-43c0-b724-43072e28fa5c/352b2627-6e8e-43c0-b724-43072e28fa5c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 1149.955201] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-923e213c-62be-4a5c-80b5-3b43d5a08f9d {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.961852] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1149.961852] env[62000]: value = "task-883015" [ 1149.961852] env[62000]: _type = "Task" [ 1149.961852] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1149.968888] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.471777] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883015, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.423391} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.472146] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72/37b135ac-1e3f-46c1-bb0e-ce7f0b69db72.vmdk to [datastore1] 352b2627-6e8e-43c0-b724-43072e28fa5c/352b2627-6e8e-43c0-b724-43072e28fa5c.vmdk {{(pid=62000) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 1150.472241] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Extending root virtual disk to 1048576 {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1150.472487] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9b79e5be-0ea8-47e2-a4a6-9fa118b06276 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.478878] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1150.478878] env[62000]: value = "task-883016" [ 1150.478878] env[62000]: _type = "Task" [ 1150.478878] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.486041] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.988295] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066797} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.988563] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Extended root virtual disk {{(pid=62000) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1150.989332] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0e6300-0d2f-4050-9820-39843a16f884 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.010328] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 352b2627-6e8e-43c0-b724-43072e28fa5c/352b2627-6e8e-43c0-b724-43072e28fa5c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1151.010545] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-380ad62d-cafc-42cb-93f9-ab4c3d0679c1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.030466] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1151.030466] env[62000]: value = "task-883017" [ 1151.030466] env[62000]: _type = "Task" [ 1151.030466] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.037701] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883017, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.540612] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883017, 'name': ReconfigVM_Task, 'duration_secs': 0.282068} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.540969] env[62000]: DEBUG nova.virt.vmwareapi.volumeops [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 352b2627-6e8e-43c0-b724-43072e28fa5c/352b2627-6e8e-43c0-b724-43072e28fa5c.vmdk or device None with type sparse {{(pid=62000) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1151.541543] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d5f89a35-b3ed-4817-8b45-19455d83e0a6 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.548066] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1151.548066] env[62000]: value = "task-883018" [ 1151.548066] env[62000]: _type = "Task" [ 1151.548066] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.555661] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883018, 'name': Rename_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.057779] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883018, 'name': Rename_Task, 'duration_secs': 0.129163} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.058083] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Powering on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 1152.058334] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1fb1b464-5f78-4531-ad51-781a067657f5 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.064741] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1152.064741] env[62000]: value = "task-883019" [ 1152.064741] env[62000]: _type = "Task" [ 1152.064741] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.071683] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.574844] env[62000]: DEBUG oslo_vmware.api [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883019, 'name': PowerOnVM_Task, 'duration_secs': 0.441421} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.575235] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Powered on the VM {{(pid=62000) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 1152.575350] env[62000]: INFO nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Took 6.77 seconds to spawn the instance on the hypervisor. [ 1152.575515] env[62000]: DEBUG nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Checking state {{(pid=62000) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1152.576281] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ba141a-b689-40c9-af9e-049fb6548867 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.093118] env[62000]: INFO nova.compute.manager [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Took 11.45 seconds to build instance. [ 1153.595429] env[62000]: DEBUG oslo_concurrency.lockutils [None req-62f4dbf0-e74c-4e94-b197-ad36e67b8ba1 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.957s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.771820] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.772251] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.772549] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "15fde663-fac4-4198-962f-8f814f1317f6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1153.772810] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1153.773084] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1153.775819] env[62000]: INFO nova.compute.manager [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Terminating instance [ 1153.777630] env[62000]: DEBUG nova.compute.manager [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1153.777919] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1153.778822] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1e46a9b-172f-4e26-b94b-31376fbaa825 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.786117] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1153.786349] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-42fc36b4-0070-4304-ad35-fbcc952f9add {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.791609] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1153.791609] env[62000]: value = "task-883020" [ 1153.791609] env[62000]: _type = "Task" [ 1153.791609] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.798961] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-883020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.065256] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "352b2627-6e8e-43c0-b724-43072e28fa5c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.065618] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.065875] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.066111] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.066331] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.068592] env[62000]: INFO nova.compute.manager [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Terminating instance [ 1154.070381] env[62000]: DEBUG nova.compute.manager [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Start destroying the instance on the hypervisor. {{(pid=62000) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1154.070617] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Destroying instance {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 1154.071482] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31e2ab3-d2f3-4fff-86ac-1b8940bc4842 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.079344] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Powering off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 1154.079580] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ab9b6b32-09d4-4234-97ea-249a94f82258 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.084999] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1154.084999] env[62000]: value = "task-883021" [ 1154.084999] env[62000]: _type = "Task" [ 1154.084999] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.092299] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883021, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.301406] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-883020, 'name': PowerOffVM_Task, 'duration_secs': 0.180585} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.301719] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.301907] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1154.302221] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-47aa0b07-52a2-4e78-93a1-dc4a4f548988 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.364163] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1154.364398] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleting contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1154.364606] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleting the datastore file [datastore2] 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.364875] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c48b00fb-88ea-47eb-a2c5-f6817ca40fbf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.371258] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for the task: (returnval){ [ 1154.371258] env[62000]: value = "task-883023" [ 1154.371258] env[62000]: _type = "Task" [ 1154.371258] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.378905] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-883023, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.597644] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883021, 'name': PowerOffVM_Task, 'duration_secs': 0.189819} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.598016] env[62000]: DEBUG nova.virt.vmwareapi.vm_util [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Powered off the VM {{(pid=62000) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 1154.598099] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Unregistering the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 1154.598361] env[62000]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5b98a7d2-da17-4dab-914e-48245f61800f {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.663791] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Unregistered the VM {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 1154.664052] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Deleting contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 1154.664249] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Deleting the datastore file [datastore1] 352b2627-6e8e-43c0-b724-43072e28fa5c {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1154.664516] env[62000]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a3ea7432-5e6f-45f5-bf41-1635bd88b372 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.670562] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for the task: (returnval){ [ 1154.670562] env[62000]: value = "task-883025" [ 1154.670562] env[62000]: _type = "Task" [ 1154.670562] env[62000]: } to complete. {{(pid=62000) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.679033] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.881230] env[62000]: DEBUG oslo_vmware.api [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Task: {'id': task-883023, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149413} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.881472] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1154.881605] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deleted contents of the VM from datastore datastore2 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1154.881783] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1154.881958] env[62000]: INFO nova.compute.manager [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1154.882266] env[62000]: DEBUG oslo.service.loopingcall [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1154.882469] env[62000]: DEBUG nova.compute.manager [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1154.882563] env[62000]: DEBUG nova.network.neutron [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1155.180332] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883025, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1155.547555] env[62000]: DEBUG nova.compute.manager [req-2c16ad0b-36ff-4887-ab38-d89fd7c184df req-8cc787ac-9d2e-4908-855b-c97a1f7b0571 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Received event network-vif-deleted-2ec45f71-6b9c-4a0a-bd0f-f73708c92053 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1155.547917] env[62000]: INFO nova.compute.manager [req-2c16ad0b-36ff-4887-ab38-d89fd7c184df req-8cc787ac-9d2e-4908-855b-c97a1f7b0571 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Neutron deleted interface 2ec45f71-6b9c-4a0a-bd0f-f73708c92053; detaching it from the instance and deleting it from the info cache [ 1155.548252] env[62000]: DEBUG nova.network.neutron [req-2c16ad0b-36ff-4887-ab38-d89fd7c184df req-8cc787ac-9d2e-4908-855b-c97a1f7b0571 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1155.680914] env[62000]: DEBUG oslo_vmware.api [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Task: {'id': task-883025, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.741653} completed successfully. {{(pid=62000) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1155.681368] env[62000]: DEBUG nova.virt.vmwareapi.ds_util [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Deleted the datastore file {{(pid=62000) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1155.681368] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Deleted contents of the VM from datastore datastore1 {{(pid=62000) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 1155.681548] env[62000]: DEBUG nova.virt.vmwareapi.vmops [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Instance destroyed {{(pid=62000) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 1155.681722] env[62000]: INFO nova.compute.manager [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1155.681952] env[62000]: DEBUG oslo.service.loopingcall [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62000) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1155.682178] env[62000]: DEBUG nova.compute.manager [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Deallocating network for instance {{(pid=62000) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1155.682275] env[62000]: DEBUG nova.network.neutron [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] deallocate_for_instance() {{(pid=62000) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 1156.017429] env[62000]: DEBUG nova.network.neutron [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.051635] env[62000]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9de73e4f-c261-43ad-acf0-0f0e88a9c2a9 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.063771] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af729266-1c64-4efd-868f-6f92f7ca5fc0 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1156.087518] env[62000]: DEBUG nova.compute.manager [req-2c16ad0b-36ff-4887-ab38-d89fd7c184df req-8cc787ac-9d2e-4908-855b-c97a1f7b0571 service nova] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Detach interface failed, port_id=2ec45f71-6b9c-4a0a-bd0f-f73708c92053, reason: Instance 15fde663-fac4-4198-962f-8f814f1317f6 could not be found. {{(pid=62000) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10963}} [ 1156.388827] env[62000]: DEBUG nova.network.neutron [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Updating instance_info_cache with network_info: [] {{(pid=62000) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.519722] env[62000]: INFO nova.compute.manager [-] [instance: 15fde663-fac4-4198-962f-8f814f1317f6] Took 1.64 seconds to deallocate network for instance. [ 1156.891986] env[62000]: INFO nova.compute.manager [-] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Took 1.21 seconds to deallocate network for instance. [ 1157.026348] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.026622] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.026845] env[62000]: DEBUG nova.objects.instance [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lazy-loading 'resources' on Instance uuid 15fde663-fac4-4198-962f-8f814f1317f6 {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1157.398335] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.571753] env[62000]: DEBUG nova.compute.manager [req-d66bf623-79ef-4bc5-a912-b14f976e8e5a req-cf841846-26c1-4d6b-bca3-3d4b7b0ffea1 service nova] [instance: 352b2627-6e8e-43c0-b724-43072e28fa5c] Received event network-vif-deleted-dc74c58b-7dae-4321-b90e-10ea9a636007 {{(pid=62000) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 1157.573817] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fa4730d-2ba9-4d40-af95-73c7e82c5d5e {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.581169] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2de31c3-cb38-4cb9-bd8e-db95ba02fe7b {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.611694] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8f439d3-ed5a-4bb5-8e09-1ec8dc4b87aa {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.618371] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9f2b6e-5cf7-44ef-9228-8b0a7a5721f1 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.630966] env[62000]: DEBUG nova.compute.provider_tree [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1158.133573] env[62000]: DEBUG nova.scheduler.client.report [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1158.638182] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.611s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1158.641557] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.243s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1158.641662] env[62000]: DEBUG nova.objects.instance [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lazy-loading 'resources' on Instance uuid 352b2627-6e8e-43c0-b724-43072e28fa5c {{(pid=62000) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1158.659998] env[62000]: INFO nova.scheduler.client.report [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Deleted allocations for instance 15fde663-fac4-4198-962f-8f814f1317f6 [ 1159.167630] env[62000]: DEBUG oslo_concurrency.lockutils [None req-f5fad2a6-352a-444b-83a2-25f6188d4bca tempest-AttachVolumeShelveTestJSON-1048694501 tempest-AttachVolumeShelveTestJSON-1048694501-project-member] Lock "15fde663-fac4-4198-962f-8f814f1317f6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.395s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1159.169410] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b1d8d4a-9ce2-417f-9329-d8b7bb49cd3a {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.177887] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d918ef32-b7f8-4f39-b818-9435b956d3f8 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.211565] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96c4c676-9f9c-4a4e-a356-e0519111b754 {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.218953] env[62000]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-281a7a2c-c3c7-4335-89a8-839ded6cd9cf {{(pid=62000) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.232408] env[62000]: DEBUG nova.compute.provider_tree [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Inventory has not changed in ProviderTree for provider: bcfccf20-49dd-4b91-819e-4373e67bf5ec {{(pid=62000) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1159.734909] env[62000]: DEBUG nova.scheduler.client.report [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Inventory has not changed for provider bcfccf20-49dd-4b91-819e-4373e67bf5ec based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 53, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62000) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1160.240071] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.599s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1160.258192] env[62000]: INFO nova.scheduler.client.report [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Deleted allocations for instance 352b2627-6e8e-43c0-b724-43072e28fa5c [ 1160.766564] env[62000]: DEBUG oslo_concurrency.lockutils [None req-67cb0824-0484-4491-bd1c-8b678a085497 tempest-ServerAddressesNegativeTestJSON-122027565 tempest-ServerAddressesNegativeTestJSON-122027565-project-member] Lock "352b2627-6e8e-43c0-b724-43072e28fa5c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.701s {{(pid=62000) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}